var/home/core/zuul-output/0000755000175000017500000000000015066762754014547 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015066776755015521 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005051650515066776746017734 0ustar rootrootSep 30 14:19:36 crc systemd[1]: Starting Kubernetes Kubelet... Sep 30 14:19:36 crc restorecon[4567]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:36 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 30 14:19:37 crc restorecon[4567]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Sep 30 14:19:38 crc kubenswrapper[4799]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 14:19:38 crc kubenswrapper[4799]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Sep 30 14:19:38 crc kubenswrapper[4799]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 14:19:38 crc kubenswrapper[4799]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 14:19:38 crc kubenswrapper[4799]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 30 14:19:38 crc kubenswrapper[4799]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.300753 4799 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304112 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304133 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304138 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304149 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304154 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304159 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304163 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304167 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304171 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304175 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304178 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304182 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304186 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304189 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304193 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304197 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304200 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304204 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304207 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304212 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304217 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304221 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304225 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304229 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304233 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304236 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304239 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304244 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304250 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304254 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304270 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304275 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304279 4799 feature_gate.go:330] unrecognized feature gate: Example Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304283 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304286 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304290 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304294 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304298 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304305 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304309 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304313 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304316 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304320 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304325 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304329 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304334 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304337 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304342 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304347 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304352 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304355 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304359 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304362 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304366 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304370 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304373 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304376 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304380 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304383 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304387 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304390 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304394 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304397 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304401 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304404 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304407 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304411 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304414 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304418 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304422 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.304426 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304492 4799 flags.go:64] FLAG: --address="0.0.0.0" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304499 4799 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304507 4799 flags.go:64] FLAG: --anonymous-auth="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304512 4799 flags.go:64] FLAG: --application-metrics-count-limit="100" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304517 4799 flags.go:64] FLAG: --authentication-token-webhook="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304521 4799 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304527 4799 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304532 4799 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304537 4799 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304542 4799 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304547 4799 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304552 4799 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304556 4799 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304560 4799 flags.go:64] FLAG: --cgroup-root="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304564 4799 flags.go:64] FLAG: --cgroups-per-qos="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304568 4799 flags.go:64] FLAG: --client-ca-file="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304572 4799 flags.go:64] FLAG: --cloud-config="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304576 4799 flags.go:64] FLAG: --cloud-provider="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304580 4799 flags.go:64] FLAG: --cluster-dns="[]" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304587 4799 flags.go:64] FLAG: --cluster-domain="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304591 4799 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304595 4799 flags.go:64] FLAG: --config-dir="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304599 4799 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304603 4799 flags.go:64] FLAG: --container-log-max-files="5" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304609 4799 flags.go:64] FLAG: --container-log-max-size="10Mi" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304613 4799 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304617 4799 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304622 4799 flags.go:64] FLAG: --containerd-namespace="k8s.io" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304626 4799 flags.go:64] FLAG: --contention-profiling="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304630 4799 flags.go:64] FLAG: --cpu-cfs-quota="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304634 4799 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304638 4799 flags.go:64] FLAG: --cpu-manager-policy="none" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304642 4799 flags.go:64] FLAG: --cpu-manager-policy-options="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304660 4799 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304664 4799 flags.go:64] FLAG: --enable-controller-attach-detach="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304668 4799 flags.go:64] FLAG: --enable-debugging-handlers="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304672 4799 flags.go:64] FLAG: --enable-load-reader="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304676 4799 flags.go:64] FLAG: --enable-server="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304681 4799 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304686 4799 flags.go:64] FLAG: --event-burst="100" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304691 4799 flags.go:64] FLAG: --event-qps="50" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304695 4799 flags.go:64] FLAG: --event-storage-age-limit="default=0" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304699 4799 flags.go:64] FLAG: --event-storage-event-limit="default=0" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304703 4799 flags.go:64] FLAG: --eviction-hard="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304708 4799 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304713 4799 flags.go:64] FLAG: --eviction-minimum-reclaim="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304717 4799 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304723 4799 flags.go:64] FLAG: --eviction-soft="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304727 4799 flags.go:64] FLAG: --eviction-soft-grace-period="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304731 4799 flags.go:64] FLAG: --exit-on-lock-contention="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304735 4799 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304739 4799 flags.go:64] FLAG: --experimental-mounter-path="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304743 4799 flags.go:64] FLAG: --fail-cgroupv1="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304747 4799 flags.go:64] FLAG: --fail-swap-on="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304751 4799 flags.go:64] FLAG: --feature-gates="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304757 4799 flags.go:64] FLAG: --file-check-frequency="20s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304762 4799 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304766 4799 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304770 4799 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304774 4799 flags.go:64] FLAG: --healthz-port="10248" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304778 4799 flags.go:64] FLAG: --help="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304783 4799 flags.go:64] FLAG: --hostname-override="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304786 4799 flags.go:64] FLAG: --housekeeping-interval="10s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304790 4799 flags.go:64] FLAG: --http-check-frequency="20s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304795 4799 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304799 4799 flags.go:64] FLAG: --image-credential-provider-config="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304803 4799 flags.go:64] FLAG: --image-gc-high-threshold="85" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304806 4799 flags.go:64] FLAG: --image-gc-low-threshold="80" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304810 4799 flags.go:64] FLAG: --image-service-endpoint="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304814 4799 flags.go:64] FLAG: --kernel-memcg-notification="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304818 4799 flags.go:64] FLAG: --kube-api-burst="100" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304822 4799 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304826 4799 flags.go:64] FLAG: --kube-api-qps="50" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304830 4799 flags.go:64] FLAG: --kube-reserved="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304834 4799 flags.go:64] FLAG: --kube-reserved-cgroup="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304838 4799 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304843 4799 flags.go:64] FLAG: --kubelet-cgroups="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304847 4799 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304851 4799 flags.go:64] FLAG: --lock-file="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304856 4799 flags.go:64] FLAG: --log-cadvisor-usage="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304860 4799 flags.go:64] FLAG: --log-flush-frequency="5s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304865 4799 flags.go:64] FLAG: --log-json-info-buffer-size="0" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304871 4799 flags.go:64] FLAG: --log-json-split-stream="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304876 4799 flags.go:64] FLAG: --log-text-info-buffer-size="0" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304879 4799 flags.go:64] FLAG: --log-text-split-stream="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304883 4799 flags.go:64] FLAG: --logging-format="text" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304888 4799 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304893 4799 flags.go:64] FLAG: --make-iptables-util-chains="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304897 4799 flags.go:64] FLAG: --manifest-url="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304901 4799 flags.go:64] FLAG: --manifest-url-header="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304906 4799 flags.go:64] FLAG: --max-housekeeping-interval="15s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304911 4799 flags.go:64] FLAG: --max-open-files="1000000" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304917 4799 flags.go:64] FLAG: --max-pods="110" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304921 4799 flags.go:64] FLAG: --maximum-dead-containers="-1" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304925 4799 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304929 4799 flags.go:64] FLAG: --memory-manager-policy="None" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304932 4799 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304937 4799 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304941 4799 flags.go:64] FLAG: --node-ip="192.168.126.11" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304944 4799 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304953 4799 flags.go:64] FLAG: --node-status-max-images="50" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304957 4799 flags.go:64] FLAG: --node-status-update-frequency="10s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304961 4799 flags.go:64] FLAG: --oom-score-adj="-999" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304965 4799 flags.go:64] FLAG: --pod-cidr="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304969 4799 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304975 4799 flags.go:64] FLAG: --pod-manifest-path="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304979 4799 flags.go:64] FLAG: --pod-max-pids="-1" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304983 4799 flags.go:64] FLAG: --pods-per-core="0" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304987 4799 flags.go:64] FLAG: --port="10250" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304991 4799 flags.go:64] FLAG: --protect-kernel-defaults="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304995 4799 flags.go:64] FLAG: --provider-id="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.304999 4799 flags.go:64] FLAG: --qos-reserved="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305003 4799 flags.go:64] FLAG: --read-only-port="10255" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305008 4799 flags.go:64] FLAG: --register-node="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305012 4799 flags.go:64] FLAG: --register-schedulable="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305015 4799 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305022 4799 flags.go:64] FLAG: --registry-burst="10" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305026 4799 flags.go:64] FLAG: --registry-qps="5" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305030 4799 flags.go:64] FLAG: --reserved-cpus="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305034 4799 flags.go:64] FLAG: --reserved-memory="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305042 4799 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305046 4799 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305050 4799 flags.go:64] FLAG: --rotate-certificates="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305054 4799 flags.go:64] FLAG: --rotate-server-certificates="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305058 4799 flags.go:64] FLAG: --runonce="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305062 4799 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305066 4799 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305071 4799 flags.go:64] FLAG: --seccomp-default="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305075 4799 flags.go:64] FLAG: --serialize-image-pulls="true" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305079 4799 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305083 4799 flags.go:64] FLAG: --storage-driver-db="cadvisor" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305087 4799 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305091 4799 flags.go:64] FLAG: --storage-driver-password="root" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305095 4799 flags.go:64] FLAG: --storage-driver-secure="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305100 4799 flags.go:64] FLAG: --storage-driver-table="stats" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305104 4799 flags.go:64] FLAG: --storage-driver-user="root" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305108 4799 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305112 4799 flags.go:64] FLAG: --sync-frequency="1m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305116 4799 flags.go:64] FLAG: --system-cgroups="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305120 4799 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305126 4799 flags.go:64] FLAG: --system-reserved-cgroup="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305129 4799 flags.go:64] FLAG: --tls-cert-file="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305133 4799 flags.go:64] FLAG: --tls-cipher-suites="[]" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305138 4799 flags.go:64] FLAG: --tls-min-version="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305143 4799 flags.go:64] FLAG: --tls-private-key-file="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305146 4799 flags.go:64] FLAG: --topology-manager-policy="none" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305151 4799 flags.go:64] FLAG: --topology-manager-policy-options="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305155 4799 flags.go:64] FLAG: --topology-manager-scope="container" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305158 4799 flags.go:64] FLAG: --v="2" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305164 4799 flags.go:64] FLAG: --version="false" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305169 4799 flags.go:64] FLAG: --vmodule="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305174 4799 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305180 4799 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305270 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305274 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305278 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305282 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305286 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305290 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305294 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305298 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305301 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305305 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305309 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305312 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305316 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305320 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305323 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305327 4799 feature_gate.go:330] unrecognized feature gate: Example Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305331 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305335 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305338 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305342 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305346 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305350 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305353 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305361 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305364 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305367 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305371 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305376 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305380 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305384 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305387 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305392 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305396 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305399 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305403 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305406 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305410 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305414 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305418 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305423 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305427 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305431 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305435 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305439 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305444 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305448 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305452 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305456 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305460 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305463 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305467 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305471 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305475 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305479 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305482 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305487 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305491 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305494 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305498 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305502 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305506 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305510 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305514 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305520 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305525 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305529 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305533 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305538 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305542 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305546 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.305550 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.305557 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.312937 4799 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.312963 4799 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313018 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313025 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313029 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313033 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313038 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313042 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313046 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313049 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313053 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313056 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313060 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313063 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313067 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313070 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313074 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313078 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313081 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313085 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313088 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313092 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313095 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313099 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313102 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313106 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313110 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313115 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313119 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313124 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313130 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313136 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313140 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313145 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313149 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313153 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313157 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313161 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313164 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313168 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313172 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313177 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313181 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313185 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313189 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313192 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313197 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313200 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313204 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313207 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313211 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313215 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313219 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313224 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313229 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313233 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313237 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313241 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313245 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313250 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313253 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313258 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313261 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313265 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313269 4799 feature_gate.go:330] unrecognized feature gate: Example Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313273 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313277 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313280 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313284 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313288 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313291 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313295 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313298 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.313304 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313406 4799 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313412 4799 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313416 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313420 4799 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313425 4799 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313430 4799 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313434 4799 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313437 4799 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313441 4799 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313445 4799 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313448 4799 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313451 4799 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313455 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313458 4799 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313462 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313465 4799 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313469 4799 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313472 4799 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313476 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313479 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313484 4799 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313487 4799 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313490 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313494 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313498 4799 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313503 4799 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313507 4799 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313511 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313515 4799 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313518 4799 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313522 4799 feature_gate.go:330] unrecognized feature gate: Example Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313527 4799 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313531 4799 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313535 4799 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313538 4799 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313542 4799 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313545 4799 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313549 4799 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313553 4799 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313556 4799 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313559 4799 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313563 4799 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313567 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313571 4799 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313575 4799 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313579 4799 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313583 4799 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313587 4799 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313591 4799 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313594 4799 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313598 4799 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313602 4799 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313605 4799 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313608 4799 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313612 4799 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313616 4799 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313619 4799 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313624 4799 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313628 4799 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313632 4799 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313636 4799 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313640 4799 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313644 4799 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313664 4799 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313669 4799 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313673 4799 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313677 4799 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313681 4799 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313685 4799 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313688 4799 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.313692 4799 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.313697 4799 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.314635 4799 server.go:940] "Client rotation is on, will bootstrap in background" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.319072 4799 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.319349 4799 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.320832 4799 server.go:997] "Starting client certificate rotation" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.320892 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.321859 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-08 10:44:24.902706374 +0000 UTC Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.321973 4799 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 932h24m46.580737115s for next certificate rotation Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.349876 4799 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.351696 4799 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.366851 4799 log.go:25] "Validated CRI v1 runtime API" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.397291 4799 log.go:25] "Validated CRI v1 image API" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.398904 4799 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.405728 4799 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-09-30-13-50-11-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.405768 4799 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.422077 4799 manager.go:217] Machine: {Timestamp:2025-09-30 14:19:38.417833025 +0000 UTC m=+0.501433492 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:86f82190-ecf6-4439-9be4-52d5a249c8f1 BootID:78226594-f56c-4f63-88aa-8c08111755bf Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:cc:2b:43 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:cc:2b:43 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b9:7f:97 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:62:a6:77 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:59:4b:9f Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:65:42:51 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ca:9e:33:b9:95:80 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ce:7f:a9:d5:b9:52 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.422268 4799 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.422400 4799 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.423713 4799 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.423895 4799 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.423926 4799 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.425381 4799 topology_manager.go:138] "Creating topology manager with none policy" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.425401 4799 container_manager_linux.go:303] "Creating device plugin manager" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.425932 4799 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.425957 4799 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.427202 4799 state_mem.go:36] "Initialized new in-memory state store" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.427291 4799 server.go:1245] "Using root directory" path="/var/lib/kubelet" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.434460 4799 kubelet.go:418] "Attempting to sync node with API server" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.434619 4799 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.434690 4799 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.434718 4799 kubelet.go:324] "Adding apiserver pod source" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.434738 4799 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.438930 4799 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.441096 4799 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.443791 4799 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.445540 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.445639 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.445698 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.445948 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.446615 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.446742 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.446868 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447219 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447446 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447502 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447548 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447598 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447692 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447752 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447810 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.447861 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.448853 4799 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.449348 4799 server.go:1280] "Started kubelet" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.450341 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.450538 4799 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.450679 4799 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 30 14:19:38 crc systemd[1]: Started Kubernetes Kubelet. Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.451508 4799 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.452434 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.452488 4799 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.452637 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 15:43:48.976058094 +0000 UTC Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.452690 4799 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1225h24m10.523370434s for next certificate rotation Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.453007 4799 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.453038 4799 volume_manager.go:287] "The desired_state_of_world populator starts" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.453057 4799 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.453115 4799 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.453467 4799 server.go:460] "Adding debug handlers to kubelet server" Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.453759 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.453827 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.454756 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="200ms" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.458641 4799 factory.go:55] Registering systemd factory Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.459875 4799 factory.go:221] Registration of the systemd container factory successfully Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.458976 4799 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.245:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a15422f6d6089 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-09-30 14:19:38.449322121 +0000 UTC m=+0.532922548,LastTimestamp:2025-09-30 14:19:38.449322121 +0000 UTC m=+0.532922548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.462114 4799 factory.go:153] Registering CRI-O factory Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.462130 4799 factory.go:221] Registration of the crio container factory successfully Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.462183 4799 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.462200 4799 factory.go:103] Registering Raw factory Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.462213 4799 manager.go:1196] Started watching for new ooms in manager Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.462881 4799 manager.go:319] Starting recovery of all containers Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.465746 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466061 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466082 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466094 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466106 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466121 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466133 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466145 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466162 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466177 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466188 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466200 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466212 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466227 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466239 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466251 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466264 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466278 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466289 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466308 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.466321 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473567 4799 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473633 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473697 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473716 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473734 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473755 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473777 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473802 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473823 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473836 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473849 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473870 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473886 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473908 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473925 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473939 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473965 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.473981 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474006 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474025 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474043 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474064 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474081 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474095 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474114 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474128 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474149 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474163 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474178 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474198 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474214 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474237 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474263 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474286 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474307 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474324 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474346 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474360 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474383 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474399 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474418 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474432 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474445 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474464 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474480 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474497 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474510 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474523 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474542 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474558 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474572 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474591 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474609 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474631 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474732 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474758 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474776 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474791 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474858 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474874 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474890 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474911 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474928 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474949 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474980 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.474997 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475019 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475034 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475055 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475072 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475086 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475107 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475126 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475143 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475164 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475180 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475197 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475213 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475228 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475248 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475265 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475285 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475300 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475315 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475342 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475367 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475384 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475405 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475427 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475443 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475464 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475483 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475498 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475519 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475534 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475552 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475574 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475590 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475604 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475618 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475636 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475669 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475689 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475703 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475719 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475737 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475749 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475767 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475781 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475798 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475816 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475830 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475858 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475872 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475885 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475904 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475918 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475935 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475949 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475962 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.475980 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476050 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476067 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476086 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476100 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476117 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476132 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476147 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476170 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476185 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476205 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476218 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476231 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476251 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476264 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476282 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476296 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476311 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476329 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476342 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476360 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476373 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476387 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476403 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476422 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476439 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476452 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476466 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476485 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476501 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476515 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476532 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476546 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476565 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476580 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476592 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476612 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476626 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476643 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476673 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476689 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476704 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476717 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476742 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476758 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476773 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476795 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476808 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476826 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476843 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476857 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476874 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476888 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476906 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476920 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476933 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476957 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476972 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.476986 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477004 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477017 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477035 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477049 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477063 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477085 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477100 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477126 4799 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477140 4799 reconstruct.go:97] "Volume reconstruction finished" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.477149 4799 reconciler.go:26] "Reconciler: start to sync state" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.479349 4799 manager.go:324] Recovery completed Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.490965 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.492895 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.492931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.492939 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.493440 4799 cpu_manager.go:225] "Starting CPU manager" policy="none" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.493482 4799 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.493502 4799 state_mem.go:36] "Initialized new in-memory state store" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.500068 4799 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.501850 4799 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.501893 4799 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.501916 4799 kubelet.go:2335] "Starting kubelet main sync loop" Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.501956 4799 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.502753 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.502817 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.518582 4799 policy_none.go:49] "None policy: Start" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.519949 4799 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.519975 4799 state_mem.go:35] "Initializing new in-memory state store" Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.554140 4799 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.561795 4799 manager.go:334] "Starting Device Plugin manager" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.561912 4799 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.561930 4799 server.go:79] "Starting device plugin registration server" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.562282 4799 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.562299 4799 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.562491 4799 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.562559 4799 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.562568 4799 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.568146 4799 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.602404 4799 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.602578 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.604092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.604138 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.604148 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.604352 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.604570 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.604631 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605202 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605252 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605437 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605579 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605636 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.605699 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.606213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.606238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.606248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.606341 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.606463 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.606493 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607544 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607593 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607552 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607567 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607851 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.607952 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.608366 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.608390 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.609233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.609249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.609257 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.609363 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.609380 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.610132 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.610165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.610178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.615281 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.615310 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.615326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.655501 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="400ms" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.662393 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.663556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.663594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.663605 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.663628 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.664192 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678585 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678639 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678687 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678709 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678736 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678760 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678787 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678814 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678883 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678939 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.678970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.679073 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.679115 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.679141 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.679192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780304 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780353 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780376 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780392 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780409 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780440 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780455 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780473 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780492 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780496 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780524 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780536 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780558 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780556 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780581 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780579 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780511 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780582 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780469 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780695 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780553 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780645 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780745 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780771 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780776 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780801 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780824 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.780802 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.865152 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.866299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.866333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.866342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.866366 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 14:19:38 crc kubenswrapper[4799]: E0930 14:19:38.866837 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.946144 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.973878 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:38 crc kubenswrapper[4799]: W0930 14:19:38.981189 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-b2e3cbe4017b34d57d0daeaf166fb674d5590451ae3df01c38055a314c093756 WatchSource:0}: Error finding container b2e3cbe4017b34d57d0daeaf166fb674d5590451ae3df01c38055a314c093756: Status 404 returned error can't find the container with id b2e3cbe4017b34d57d0daeaf166fb674d5590451ae3df01c38055a314c093756 Sep 30 14:19:38 crc kubenswrapper[4799]: I0930 14:19:38.993612 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:39 crc kubenswrapper[4799]: W0930 14:19:39.006753 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-bc06ca28c856a99a53e99c0daca004f06789eab039af9c3ed7ebaf0f6060ed24 WatchSource:0}: Error finding container bc06ca28c856a99a53e99c0daca004f06789eab039af9c3ed7ebaf0f6060ed24: Status 404 returned error can't find the container with id bc06ca28c856a99a53e99c0daca004f06789eab039af9c3ed7ebaf0f6060ed24 Sep 30 14:19:39 crc kubenswrapper[4799]: W0930 14:19:39.015237 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-849dac29b4d4c250e80ecd0c69973c2719df2ec01723ffd0a7b80a4fc2154596 WatchSource:0}: Error finding container 849dac29b4d4c250e80ecd0c69973c2719df2ec01723ffd0a7b80a4fc2154596: Status 404 returned error can't find the container with id 849dac29b4d4c250e80ecd0c69973c2719df2ec01723ffd0a7b80a4fc2154596 Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.018739 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.024126 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 30 14:19:39 crc kubenswrapper[4799]: W0930 14:19:39.028845 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c76de09578f1ac1011aa7f531e9fbbec2da3ab1ec4ee508a85499e96d6bb2752 WatchSource:0}: Error finding container c76de09578f1ac1011aa7f531e9fbbec2da3ab1ec4ee508a85499e96d6bb2752: Status 404 returned error can't find the container with id c76de09578f1ac1011aa7f531e9fbbec2da3ab1ec4ee508a85499e96d6bb2752 Sep 30 14:19:39 crc kubenswrapper[4799]: W0930 14:19:39.038606 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8fb46e34e21794079c6353aab67e0e0ff65f75fb5055a59a60de6ad509aa6959 WatchSource:0}: Error finding container 8fb46e34e21794079c6353aab67e0e0ff65f75fb5055a59a60de6ad509aa6959: Status 404 returned error can't find the container with id 8fb46e34e21794079c6353aab67e0e0ff65f75fb5055a59a60de6ad509aa6959 Sep 30 14:19:39 crc kubenswrapper[4799]: E0930 14:19:39.056510 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="800ms" Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.267101 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.268902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.268945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.268956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.268990 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 14:19:39 crc kubenswrapper[4799]: E0930 14:19:39.269458 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Sep 30 14:19:39 crc kubenswrapper[4799]: W0930 14:19:39.363811 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:39 crc kubenswrapper[4799]: E0930 14:19:39.363970 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:39 crc kubenswrapper[4799]: W0930 14:19:39.445821 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:39 crc kubenswrapper[4799]: E0930 14:19:39.445898 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.451549 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.506916 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b2e3cbe4017b34d57d0daeaf166fb674d5590451ae3df01c38055a314c093756"} Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.507978 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8fb46e34e21794079c6353aab67e0e0ff65f75fb5055a59a60de6ad509aa6959"} Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.508909 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c76de09578f1ac1011aa7f531e9fbbec2da3ab1ec4ee508a85499e96d6bb2752"} Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.509750 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"849dac29b4d4c250e80ecd0c69973c2719df2ec01723ffd0a7b80a4fc2154596"} Sep 30 14:19:39 crc kubenswrapper[4799]: I0930 14:19:39.511051 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bc06ca28c856a99a53e99c0daca004f06789eab039af9c3ed7ebaf0f6060ed24"} Sep 30 14:19:39 crc kubenswrapper[4799]: W0930 14:19:39.709363 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:39 crc kubenswrapper[4799]: E0930 14:19:39.709452 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:39 crc kubenswrapper[4799]: E0930 14:19:39.857257 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="1.6s" Sep 30 14:19:39 crc kubenswrapper[4799]: W0930 14:19:39.906776 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:39 crc kubenswrapper[4799]: E0930 14:19:39.906843 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.070521 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.071843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.071874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.071887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.071914 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 14:19:40 crc kubenswrapper[4799]: E0930 14:19:40.072356 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.451112 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.514552 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00" exitCode=0 Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.514621 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00"} Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.514700 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.515548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.515578 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.515586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.519678 4799 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="05566b59254df7f2aa21029d8333078db94541ab7e274195010613cad824e85e" exitCode=0 Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.519866 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.520047 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"05566b59254df7f2aa21029d8333078db94541ab7e274195010613cad824e85e"} Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.520791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.520826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.520839 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.523413 4799 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286" exitCode=0 Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.523472 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.523525 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286"} Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.524817 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.524841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.524866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.527790 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.528092 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063"} Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.528118 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803"} Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.528129 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265"} Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.528138 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca"} Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.528556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.528598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.528610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.535326 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e" exitCode=0 Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.535372 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e"} Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.535521 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.536560 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.536592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.536603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.542566 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.544191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.544236 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:40 crc kubenswrapper[4799]: I0930 14:19:40.544250 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.451586 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:41 crc kubenswrapper[4799]: E0930 14:19:41.458555 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.245:6443: connect: connection refused" interval="3.2s" Sep 30 14:19:41 crc kubenswrapper[4799]: W0930 14:19:41.502321 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:41 crc kubenswrapper[4799]: E0930 14:19:41.502408 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.542735 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.543009 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.543027 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.543128 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.544429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.544463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.544475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.547456 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4acefa926079610a5207a5631f12848fef37a45f1d1739757ce38154be9480ad"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.547522 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.547537 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.547552 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.547562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.547572 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.548577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.548609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.548623 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.549333 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be" exitCode=0 Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.549400 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.549438 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.550156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.550184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.550193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.551708 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"31b21f4329ffe89a3a4fa23f7fd3d5168eebc20e1849632375c40710b11abc80"} Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.551795 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.552610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.552635 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.552659 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.553020 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.553771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.553798 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.553812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:41 crc kubenswrapper[4799]: W0930 14:19:41.576148 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.245:6443: connect: connection refused Sep 30 14:19:41 crc kubenswrapper[4799]: E0930 14:19:41.576256 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.245:6443: connect: connection refused" logger="UnhandledError" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.672538 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.673837 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.673866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.673874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.673896 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 14:19:41 crc kubenswrapper[4799]: E0930 14:19:41.674328 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.245:6443: connect: connection refused" node="crc" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.676346 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:41 crc kubenswrapper[4799]: I0930 14:19:41.907944 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.556232 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.559318 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4acefa926079610a5207a5631f12848fef37a45f1d1739757ce38154be9480ad" exitCode=255 Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.559373 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4acefa926079610a5207a5631f12848fef37a45f1d1739757ce38154be9480ad"} Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.559437 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.560455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.560525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.560541 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.561122 4799 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835" exitCode=0 Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.561197 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.561232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835"} Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.561248 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.561202 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.561269 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.561288 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.561730 4799 scope.go:117] "RemoveContainer" containerID="4acefa926079610a5207a5631f12848fef37a45f1d1739757ce38154be9480ad" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.562332 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.562353 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.562362 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.563126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.563148 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.563158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.563615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.563636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.563667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.564034 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.564052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.564059 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:42 crc kubenswrapper[4799]: I0930 14:19:42.749497 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.568116 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.570439 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c"} Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.570539 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.570620 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.571855 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.571908 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.571926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.575555 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c"} Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.575615 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2"} Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.575644 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b"} Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.575699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e"} Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.575725 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7"} Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.575666 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.575793 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.575872 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.577154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.577194 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.577222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.577509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.577535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.577549 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.578277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.578334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:43 crc kubenswrapper[4799]: I0930 14:19:43.578358 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.578430 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.578474 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.578439 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.579602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.579627 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.579636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.580441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.580509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.580538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.617451 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.875115 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.876921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.876963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.876972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:44 crc kubenswrapper[4799]: I0930 14:19:44.877004 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 14:19:45 crc kubenswrapper[4799]: I0930 14:19:45.107864 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:45 crc kubenswrapper[4799]: I0930 14:19:45.212577 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:45 crc kubenswrapper[4799]: I0930 14:19:45.581489 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:45 crc kubenswrapper[4799]: I0930 14:19:45.582768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:45 crc kubenswrapper[4799]: I0930 14:19:45.582811 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:45 crc kubenswrapper[4799]: I0930 14:19:45.582830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:46 crc kubenswrapper[4799]: I0930 14:19:46.585137 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:46 crc kubenswrapper[4799]: I0930 14:19:46.586284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:46 crc kubenswrapper[4799]: I0930 14:19:46.586323 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:46 crc kubenswrapper[4799]: I0930 14:19:46.586333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:47 crc kubenswrapper[4799]: I0930 14:19:47.743754 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:47 crc kubenswrapper[4799]: I0930 14:19:47.744094 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:47 crc kubenswrapper[4799]: I0930 14:19:47.746301 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:47 crc kubenswrapper[4799]: I0930 14:19:47.746408 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:47 crc kubenswrapper[4799]: I0930 14:19:47.746420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:47 crc kubenswrapper[4799]: I0930 14:19:47.753891 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.202033 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.202369 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.204640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.204782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.204818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:48 crc kubenswrapper[4799]: E0930 14:19:48.568245 4799 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.589713 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.590800 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.590854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:48 crc kubenswrapper[4799]: I0930 14:19:48.590868 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:49 crc kubenswrapper[4799]: I0930 14:19:49.846551 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:49 crc kubenswrapper[4799]: I0930 14:19:49.846954 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:49 crc kubenswrapper[4799]: I0930 14:19:49.849151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:49 crc kubenswrapper[4799]: I0930 14:19:49.849234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:49 crc kubenswrapper[4799]: I0930 14:19:49.849272 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:49 crc kubenswrapper[4799]: I0930 14:19:49.851588 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:50 crc kubenswrapper[4799]: I0930 14:19:50.594912 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:50 crc kubenswrapper[4799]: I0930 14:19:50.595736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:50 crc kubenswrapper[4799]: I0930 14:19:50.595763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:50 crc kubenswrapper[4799]: I0930 14:19:50.595777 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:51 crc kubenswrapper[4799]: I0930 14:19:51.006274 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Sep 30 14:19:51 crc kubenswrapper[4799]: I0930 14:19:51.006603 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:51 crc kubenswrapper[4799]: I0930 14:19:51.008794 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:51 crc kubenswrapper[4799]: I0930 14:19:51.008854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:51 crc kubenswrapper[4799]: I0930 14:19:51.008867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:52 crc kubenswrapper[4799]: W0930 14:19:52.218839 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.218941 4799 trace.go:236] Trace[448552025]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 14:19:42.217) (total time: 10001ms): Sep 30 14:19:52 crc kubenswrapper[4799]: Trace[448552025]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:19:52.218) Sep 30 14:19:52 crc kubenswrapper[4799]: Trace[448552025]: [10.001415172s] [10.001415172s] END Sep 30 14:19:52 crc kubenswrapper[4799]: E0930 14:19:52.218969 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.452862 4799 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Sep 30 14:19:52 crc kubenswrapper[4799]: W0930 14:19:52.545493 4799 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.545580 4799 trace.go:236] Trace[913489735]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 14:19:42.543) (total time: 10001ms): Sep 30 14:19:52 crc kubenswrapper[4799]: Trace[913489735]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:19:52.545) Sep 30 14:19:52 crc kubenswrapper[4799]: Trace[913489735]: [10.001703719s] [10.001703719s] END Sep 30 14:19:52 crc kubenswrapper[4799]: E0930 14:19:52.545606 4799 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.599918 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.600016 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.609528 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.609596 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.847015 4799 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 30 14:19:52 crc kubenswrapper[4799]: I0930 14:19:52.847111 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 30 14:19:54 crc kubenswrapper[4799]: I0930 14:19:54.621642 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:54 crc kubenswrapper[4799]: I0930 14:19:54.621834 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:54 crc kubenswrapper[4799]: I0930 14:19:54.622075 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Sep 30 14:19:54 crc kubenswrapper[4799]: I0930 14:19:54.622111 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Sep 30 14:19:54 crc kubenswrapper[4799]: I0930 14:19:54.622733 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:54 crc kubenswrapper[4799]: I0930 14:19:54.622764 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:54 crc kubenswrapper[4799]: I0930 14:19:54.622774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:54 crc kubenswrapper[4799]: I0930 14:19:54.625213 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:55 crc kubenswrapper[4799]: I0930 14:19:55.213283 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Sep 30 14:19:55 crc kubenswrapper[4799]: I0930 14:19:55.213350 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Sep 30 14:19:55 crc kubenswrapper[4799]: I0930 14:19:55.608798 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:55 crc kubenswrapper[4799]: I0930 14:19:55.609508 4799 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Sep 30 14:19:55 crc kubenswrapper[4799]: I0930 14:19:55.609628 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Sep 30 14:19:55 crc kubenswrapper[4799]: I0930 14:19:55.610054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:55 crc kubenswrapper[4799]: I0930 14:19:55.610093 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:55 crc kubenswrapper[4799]: I0930 14:19:55.610102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:57 crc kubenswrapper[4799]: E0930 14:19:57.602788 4799 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Sep 30 14:19:57 crc kubenswrapper[4799]: I0930 14:19:57.605763 4799 trace.go:236] Trace[2002592206]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 14:19:46.654) (total time: 10950ms): Sep 30 14:19:57 crc kubenswrapper[4799]: Trace[2002592206]: ---"Objects listed" error: 10950ms (14:19:57.605) Sep 30 14:19:57 crc kubenswrapper[4799]: Trace[2002592206]: [10.950879011s] [10.950879011s] END Sep 30 14:19:57 crc kubenswrapper[4799]: I0930 14:19:57.605787 4799 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Sep 30 14:19:57 crc kubenswrapper[4799]: E0930 14:19:57.607710 4799 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Sep 30 14:19:57 crc kubenswrapper[4799]: I0930 14:19:57.607904 4799 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Sep 30 14:19:57 crc kubenswrapper[4799]: I0930 14:19:57.608879 4799 trace.go:236] Trace[693074057]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (30-Sep-2025 14:19:47.341) (total time: 10267ms): Sep 30 14:19:57 crc kubenswrapper[4799]: Trace[693074057]: ---"Objects listed" error: 10267ms (14:19:57.608) Sep 30 14:19:57 crc kubenswrapper[4799]: Trace[693074057]: [10.267346248s] [10.267346248s] END Sep 30 14:19:57 crc kubenswrapper[4799]: I0930 14:19:57.608905 4799 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.208727 4799 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Sep 30 14:19:58 crc kubenswrapper[4799]: E0930 14:19:58.569302 4799 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.616226 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.616690 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.618080 4799 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c" exitCode=255 Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.618120 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c"} Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.618203 4799 scope.go:117] "RemoveContainer" containerID="4acefa926079610a5207a5631f12848fef37a45f1d1739757ce38154be9480ad" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.618340 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.619108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.619147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.619159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.619838 4799 scope.go:117] "RemoveContainer" containerID="c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c" Sep 30 14:19:58 crc kubenswrapper[4799]: E0930 14:19:58.620034 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 14:19:58 crc kubenswrapper[4799]: I0930 14:19:58.817913 4799 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.310494 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.446800 4799 apiserver.go:52] "Watching apiserver" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.455036 4799 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.455465 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-machine-config-operator/machine-config-daemon-wr2vb","openshift-multus/multus-additional-cni-plugins-hstq9","openshift-multus/multus-sjjm8","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-dns/node-resolver-hlg57"] Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.455841 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.456002 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.456014 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.456498 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.456573 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.456601 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.455911 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.456710 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.456769 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hlg57" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.456824 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.457061 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.457428 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.457500 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.459052 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.459400 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.461957 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462020 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462058 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462087 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462030 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462182 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462226 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462181 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462302 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462373 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462492 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462768 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.462894 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.463034 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.463136 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.463257 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.463409 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.463422 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.463446 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.463676 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.463831 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.464014 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.478639 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.487924 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.497564 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.509716 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.518067 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.528040 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.536753 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.545197 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.550696 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6bhnh"] Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.551452 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.553304 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.553511 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.559798 4799 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.560230 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.560267 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.560230 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.560230 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.560610 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.560835 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.570141 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.580612 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.589233 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.599752 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.608210 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619413 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619456 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619481 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619504 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619526 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619546 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619568 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619590 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619609 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619634 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619687 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619708 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619728 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619751 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619771 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619791 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619784 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619813 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619836 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619859 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619879 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619901 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619924 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619945 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619964 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619965 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.619985 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620011 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620033 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620065 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620088 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620094 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620110 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620133 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620155 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620176 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620198 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620220 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620240 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620260 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620280 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620291 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620314 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620335 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620355 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620378 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620399 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620417 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620433 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620443 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620468 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620493 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620515 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620535 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620557 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620578 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620606 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620630 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620673 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620694 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620715 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620764 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620789 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620810 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620834 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620858 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620879 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620900 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620923 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620945 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620968 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620991 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621015 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621041 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621063 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621085 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621106 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621136 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621160 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621289 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621319 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621342 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621365 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621390 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621415 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621438 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621462 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621486 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621508 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621532 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621555 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621578 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621601 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621627 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621671 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621696 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621720 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621746 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621774 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621797 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621820 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621846 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621873 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621896 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621921 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621946 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621968 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621997 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622025 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622057 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622082 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622108 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622132 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622154 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622177 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622202 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622227 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622253 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622276 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622299 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622322 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622346 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622371 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622396 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622423 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622472 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622549 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622575 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622600 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622625 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622670 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622695 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622724 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622751 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622774 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622799 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622822 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622844 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622869 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622892 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622915 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622937 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622958 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622983 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623098 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623125 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623146 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623170 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623199 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623223 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623245 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623272 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623294 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623317 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623341 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623362 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623387 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623411 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623434 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623459 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623482 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623517 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623541 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623565 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623589 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623613 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623635 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623676 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623701 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623726 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623750 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623772 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623795 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623818 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623841 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623864 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623888 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623914 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623937 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623959 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623985 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624007 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624033 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624057 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624080 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624104 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624126 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624149 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624175 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624199 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624222 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624247 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624271 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624294 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624317 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624340 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624368 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624457 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-os-release\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624485 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624508 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-script-lib\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-cnibin\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624562 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624586 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-systemd\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624607 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cni-binary-copy\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624677 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-env-overrides\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624699 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-netns\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624722 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-etc-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624744 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-netd\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624766 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-config\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624794 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624820 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-kubelet\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624877 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-ovn-kubernetes\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624901 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-system-cni-dir\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cnibin\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624978 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625006 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625030 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625053 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-conf-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625075 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-system-cni-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625097 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-hostroot\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625123 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwl67\" (UniqueName: \"kubernetes.io/projected/3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d-kube-api-access-lwl67\") pod \"node-resolver-hlg57\" (UID: \"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\") " pod="openshift-dns/node-resolver-hlg57" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625145 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625168 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-bin\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625248 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-daemon-config\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625277 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cad3b181-b2f5-4961-8134-f471bff50c40-rootfs\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625301 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-var-lib-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625325 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-cni-bin\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625346 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-cni-multus\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625370 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cad3b181-b2f5-4961-8134-f471bff50c40-proxy-tls\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-multus-certs\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625414 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5tsx\" (UniqueName: \"kubernetes.io/projected/5517db4f-9237-4b92-a672-6fa7afcd7e3f-kube-api-access-g5tsx\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-log-socket\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625456 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-cni-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625478 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625498 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-etc-kubernetes\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625521 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g747d\" (UniqueName: \"kubernetes.io/projected/cad3b181-b2f5-4961-8134-f471bff50c40-kube-api-access-g747d\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625543 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh9hh\" (UniqueName: \"kubernetes.io/projected/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-kube-api-access-gh9hh\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625567 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625591 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625615 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-k8s-cni-cncf-io\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625666 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625690 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-netns\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-ovn\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625731 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5517db4f-9237-4b92-a672-6fa7afcd7e3f-cni-binary-copy\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625754 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-slash\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625774 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-os-release\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625798 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cad3b181-b2f5-4961-8134-f471bff50c40-mcd-auth-proxy-config\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625820 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625848 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625872 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-systemd-units\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626039 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-kubelet\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626064 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgb7m\" (UniqueName: \"kubernetes.io/projected/cce844e9-25b0-49e7-9ca3-899b986d0f6f-kube-api-access-hgb7m\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626088 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d-hosts-file\") pod \"node-resolver-hlg57\" (UID: \"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\") " pod="openshift-dns/node-resolver-hlg57" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626543 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovn-node-metrics-cert\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626576 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-socket-dir-parent\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626602 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-node-log\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626634 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626678 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626749 4799 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626767 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626782 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626798 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626813 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620439 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620705 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.620801 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621008 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621044 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621089 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621136 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621184 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621201 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621329 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621341 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621517 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621525 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621791 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621858 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621947 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.621983 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622200 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622285 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622280 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622315 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622361 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622372 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622459 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622555 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622560 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622667 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622761 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622860 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.622967 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623028 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.623782 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624204 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624403 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624595 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624778 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.624795 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.625148 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626141 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626427 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626580 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.626227 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.627682 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.627806 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.628049 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.628341 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.628449 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.628984 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.629078 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.629169 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.629272 4799 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.629934 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.630067 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.630153 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.630176 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.630373 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.630418 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.630465 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.630527 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.630640 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.631131 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.631336 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.631460 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.631605 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.631672 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.631848 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.631998 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.632137 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.632820 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.633130 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.633233 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.633603 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.633682 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.633841 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.633853 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.634019 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.634037 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.634139 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:00.134093436 +0000 UTC m=+22.217693863 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.634183 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.634684 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.634801 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635050 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635149 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.635204 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:20:00.135179357 +0000 UTC m=+22.218779784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635306 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635332 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635444 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635492 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635794 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635826 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.635851 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.636015 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.636111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.636228 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.636342 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.636499 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.636834 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.637066 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.637315 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.637776 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.637908 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.638277 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.638332 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.638632 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.638817 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.639011 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.639420 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.639599 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.640090 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.640149 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.640447 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.640524 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.640678 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.643467 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.643554 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.640766 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.640984 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.641123 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.643766 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.640706 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.641618 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.642557 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.642886 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.642999 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.641270 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.644315 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.644449 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.644813 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.644596 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.644688 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.644734 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.644800 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.644914 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.645065 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.645107 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.645507 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.645606 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.645823 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.645880 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.646002 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.638053 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.645869 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.646379 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.646412 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.646529 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.645891 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.646842 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.646849 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.647044 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.647769 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.647821 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.647981 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.648749 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.649220 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.651542 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.651957 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.652413 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.652852 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.657667 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.658616 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.659242 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.659924 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.660479 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.660547 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.660906 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.660930 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.661075 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.661191 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.661280 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.661455 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.661457 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.661628 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.661711 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.661855 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.662030 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.662152 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.662167 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.662874 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.663087 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.663248 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.663568 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.663661 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.668034 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.668388 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.668580 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.668675 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:00.16863386 +0000 UTC m=+22.252234357 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.668898 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.668974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.669455 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.669475 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.669553 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.669578 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.671763 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.671797 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.671811 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.671878 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:00.171850843 +0000 UTC m=+22.255451270 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.675985 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.681136 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.681404 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.683318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.684999 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.685276 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.685850 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.688871 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.689104 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.694238 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.694417 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.694437 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.694452 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.694508 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:00.194487841 +0000 UTC m=+22.278088268 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.695144 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.706031 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.713229 4799 scope.go:117] "RemoveContainer" containerID="c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c" Sep 30 14:19:59 crc kubenswrapper[4799]: E0930 14:19:59.713413 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.716329 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.721513 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.723366 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728063 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwl67\" (UniqueName: \"kubernetes.io/projected/3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d-kube-api-access-lwl67\") pod \"node-resolver-hlg57\" (UID: \"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\") " pod="openshift-dns/node-resolver-hlg57" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728109 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728129 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-bin\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728150 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-system-cni-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728165 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-hostroot\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cad3b181-b2f5-4961-8134-f471bff50c40-rootfs\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728194 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-var-lib-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728208 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-cni-bin\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-cni-multus\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728236 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-daemon-config\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728251 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cad3b181-b2f5-4961-8134-f471bff50c40-proxy-tls\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728264 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-multus-certs\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728279 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5tsx\" (UniqueName: \"kubernetes.io/projected/5517db4f-9237-4b92-a672-6fa7afcd7e3f-kube-api-access-g5tsx\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728294 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-log-socket\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728307 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-cni-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728322 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728335 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-etc-kubernetes\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728349 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g747d\" (UniqueName: \"kubernetes.io/projected/cad3b181-b2f5-4961-8134-f471bff50c40-kube-api-access-g747d\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728366 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh9hh\" (UniqueName: \"kubernetes.io/projected/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-kube-api-access-gh9hh\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728410 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-k8s-cni-cncf-io\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728425 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-netns\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728450 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-ovn\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5517db4f-9237-4b92-a672-6fa7afcd7e3f-cni-binary-copy\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728481 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cad3b181-b2f5-4961-8134-f471bff50c40-mcd-auth-proxy-config\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728501 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728516 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-systemd-units\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728530 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-slash\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728550 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-os-release\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728566 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgb7m\" (UniqueName: \"kubernetes.io/projected/cce844e9-25b0-49e7-9ca3-899b986d0f6f-kube-api-access-hgb7m\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728579 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d-hosts-file\") pod \"node-resolver-hlg57\" (UID: \"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\") " pod="openshift-dns/node-resolver-hlg57" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728596 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovn-node-metrics-cert\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728609 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-socket-dir-parent\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728622 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-kubelet\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728638 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-node-log\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728682 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-os-release\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728706 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728725 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-script-lib\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728741 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-cnibin\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728759 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-systemd\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728778 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cni-binary-copy\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728774 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-etc-kubernetes\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728806 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-env-overrides\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728824 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-netns\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728843 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-etc-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728864 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-netd\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728883 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-config\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728901 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-ovn-kubernetes\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728916 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-system-cni-dir\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cnibin\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728949 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-kubelet\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728976 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-conf-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.728968 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729044 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729057 4799 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729067 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729076 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729085 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729094 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729103 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729111 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729121 4799 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729129 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729138 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729146 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729154 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729164 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729173 4799 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729182 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729191 4799 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729199 4799 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729207 4799 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729216 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729224 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729232 4799 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729242 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729251 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729259 4799 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729268 4799 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729275 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729284 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729292 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729301 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729309 4799 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729317 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729325 4799 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729334 4799 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729343 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729352 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729353 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729360 4799 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-conf-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729400 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729423 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-bin\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729433 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729443 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729456 4799 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729466 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729476 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729490 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729502 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729503 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-hostroot\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729512 4799 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729477 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-system-cni-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729530 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729548 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729562 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729577 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729589 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729604 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729616 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729629 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729644 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729677 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729691 4799 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729711 4799 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729723 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729734 4799 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732992 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733015 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733028 4799 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733041 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733056 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733067 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733079 4799 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733090 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733100 4799 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733111 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733122 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733133 4799 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733146 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733157 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733169 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733181 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733192 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733203 4799 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733214 4799 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733226 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733240 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733252 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733265 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733276 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733288 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733300 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733312 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733323 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733335 4799 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733348 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733359 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733370 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733383 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733395 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733408 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733418 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733428 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733439 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733451 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733462 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733473 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733483 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733497 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733508 4799 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733521 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733532 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733544 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733556 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733570 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733582 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733593 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733606 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733619 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733633 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733666 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733680 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733692 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733703 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733718 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733730 4799 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733740 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733752 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733763 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733775 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733788 4799 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733800 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733811 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733824 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733835 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733847 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733857 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733867 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733877 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733888 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733900 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733912 4799 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733924 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733935 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733946 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733960 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733972 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.733984 4799 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734021 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734035 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734050 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734061 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734074 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734087 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734099 4799 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734110 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734123 4799 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734135 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734147 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734159 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734171 4799 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734183 4799 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734195 4799 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734207 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734218 4799 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734230 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734242 4799 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734253 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734265 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734278 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734289 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734301 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734313 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734325 4799 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734336 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734347 4799 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734360 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734371 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734383 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734394 4799 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734405 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734417 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734430 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734441 4799 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734455 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734466 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734478 4799 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734488 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734499 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734511 4799 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734523 4799 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734534 4799 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734565 4799 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734576 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.734587 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732954 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-config\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729897 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729932 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-os-release\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729934 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-node-log\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729968 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.729997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-k8s-cni-cncf-io\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.730021 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-netns\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.730052 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-ovn\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.730750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.730999 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5517db4f-9237-4b92-a672-6fa7afcd7e3f-cni-binary-copy\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731325 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-script-lib\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731353 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-cnibin\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731369 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-systemd\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731589 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cad3b181-b2f5-4961-8134-f471bff50c40-mcd-auth-proxy-config\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731717 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-cni-bin\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731741 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cad3b181-b2f5-4961-8134-f471bff50c40-rootfs\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731736 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-cni-multus\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-var-lib-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-log-socket\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-multus-certs\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731916 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-slash\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.731931 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-systemd-units\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732001 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-os-release\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732133 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-socket-dir-parent\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732146 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732176 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-var-lib-kubelet\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732201 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732235 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-cni-dir\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732255 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-system-cni-dir\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732272 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-ovn-kubernetes\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732297 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d-hosts-file\") pod \"node-resolver-hlg57\" (UID: \"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\") " pod="openshift-dns/node-resolver-hlg57" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732314 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cnibin\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732332 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5517db4f-9237-4b92-a672-6fa7afcd7e3f-host-run-netns\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732342 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5517db4f-9237-4b92-a672-6fa7afcd7e3f-multus-daemon-config\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732370 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-kubelet\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732389 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-etc-openvswitch\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732406 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-netd\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732724 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-env-overrides\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.732934 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cce844e9-25b0-49e7-9ca3-899b986d0f6f-cni-binary-copy\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.736364 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cad3b181-b2f5-4961-8134-f471bff50c40-proxy-tls\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.743163 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovn-node-metrics-cert\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.749160 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g747d\" (UniqueName: \"kubernetes.io/projected/cad3b181-b2f5-4961-8134-f471bff50c40-kube-api-access-g747d\") pod \"machine-config-daemon-wr2vb\" (UID: \"cad3b181-b2f5-4961-8134-f471bff50c40\") " pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.755506 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.755946 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5tsx\" (UniqueName: \"kubernetes.io/projected/5517db4f-9237-4b92-a672-6fa7afcd7e3f-kube-api-access-g5tsx\") pod \"multus-sjjm8\" (UID: \"5517db4f-9237-4b92-a672-6fa7afcd7e3f\") " pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.764732 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwl67\" (UniqueName: \"kubernetes.io/projected/3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d-kube-api-access-lwl67\") pod \"node-resolver-hlg57\" (UID: \"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\") " pod="openshift-dns/node-resolver-hlg57" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.766300 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgb7m\" (UniqueName: \"kubernetes.io/projected/cce844e9-25b0-49e7-9ca3-899b986d0f6f-kube-api-access-hgb7m\") pod \"multus-additional-cni-plugins-hstq9\" (UID: \"cce844e9-25b0-49e7-9ca3-899b986d0f6f\") " pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.770612 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh9hh\" (UniqueName: \"kubernetes.io/projected/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-kube-api-access-gh9hh\") pod \"ovnkube-node-6bhnh\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.774840 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.776060 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.781112 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.784141 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.788453 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.795263 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.800359 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.804137 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hstq9" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.811222 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-sjjm8" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.815147 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.817950 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hlg57" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.825883 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.834152 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: W0930 14:19:59.839901 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5517db4f_9237_4b92_a672_6fa7afcd7e3f.slice/crio-6cfee8ff0e1f44f10d455e6cb923cad425edf80f7e992d22cfcf02287108070b WatchSource:0}: Error finding container 6cfee8ff0e1f44f10d455e6cb923cad425edf80f7e992d22cfcf02287108070b: Status 404 returned error can't find the container with id 6cfee8ff0e1f44f10d455e6cb923cad425edf80f7e992d22cfcf02287108070b Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.844951 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.857125 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.864161 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.865498 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.866810 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.871820 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.886001 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.904965 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.920745 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.931252 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.942447 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.957617 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.970187 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: W0930 14:19:59.978317 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcc6c6df_d8b4_4d81_9aa5_8bc35533304b.slice/crio-944a76bbfb120710fbc9a2e94267336fa84172bc5b80c5b2eababbf3442fbd74 WatchSource:0}: Error finding container 944a76bbfb120710fbc9a2e94267336fa84172bc5b80c5b2eababbf3442fbd74: Status 404 returned error can't find the container with id 944a76bbfb120710fbc9a2e94267336fa84172bc5b80c5b2eababbf3442fbd74 Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.983096 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:19:59 crc kubenswrapper[4799]: I0930 14:19:59.995929 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.005680 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.016784 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.028354 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.043270 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.056972 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.081135 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.117095 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.132936 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.137280 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.137390 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:20:01.13737022 +0000 UTC m=+23.220970647 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.137537 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.137684 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.137739 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:01.137731471 +0000 UTC m=+23.221331898 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.154270 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.170629 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.192302 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.225566 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.238297 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.238362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.238393 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238513 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238567 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238607 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238621 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238578 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:01.238558801 +0000 UTC m=+23.322159228 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238714 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238732 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238738 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:01.238712095 +0000 UTC m=+23.322312712 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238747 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.238791 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:01.238779067 +0000 UTC m=+23.322379724 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.502676 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.502834 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.507241 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.507761 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.508514 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.509192 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.510051 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.510595 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.511196 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.511757 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.512549 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.513142 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.513667 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.514402 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.517022 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.517583 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.518538 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.519111 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.519737 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.520857 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.521390 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.521992 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.523067 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.523684 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.524589 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.525243 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.525738 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.526979 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.528013 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.528526 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.529202 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.530204 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.530772 4799 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.530889 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.533209 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.533799 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.534342 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.536008 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.537197 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.537786 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.538908 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.539744 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.540733 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.541439 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.542719 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.543700 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.544165 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.545077 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.545585 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.546874 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.547511 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.548006 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.548959 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.549618 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.551031 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.551546 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.666388 4799 generic.go:334] "Generic (PLEG): container finished" podID="cce844e9-25b0-49e7-9ca3-899b986d0f6f" containerID="d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f" exitCode=0 Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.666471 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" event={"ID":"cce844e9-25b0-49e7-9ca3-899b986d0f6f","Type":"ContainerDied","Data":"d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.666526 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" event={"ID":"cce844e9-25b0-49e7-9ca3-899b986d0f6f","Type":"ContainerStarted","Data":"ae2efb63576d0e0ea2bc9d6e607d77ab8a46e8750f0e36bb077f1daf7d33a20f"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.668247 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.668274 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.668284 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"74a94bd614fb5ccaa640f19c8e800af517792c8b1be5432d103f300770f5f66a"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.669663 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff" exitCode=0 Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.669673 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.669752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"944a76bbfb120710fbc9a2e94267336fa84172bc5b80c5b2eababbf3442fbd74"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.671686 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hlg57" event={"ID":"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d","Type":"ContainerStarted","Data":"4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.671720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hlg57" event={"ID":"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d","Type":"ContainerStarted","Data":"810ae102c7dceae7e8dcdafcc2923f145a3dbf6e44ba42c5426c07f49b7c5213"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.676180 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sjjm8" event={"ID":"5517db4f-9237-4b92-a672-6fa7afcd7e3f","Type":"ContainerStarted","Data":"f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.676224 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sjjm8" event={"ID":"5517db4f-9237-4b92-a672-6fa7afcd7e3f","Type":"ContainerStarted","Data":"6cfee8ff0e1f44f10d455e6cb923cad425edf80f7e992d22cfcf02287108070b"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.677538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"04c81764ed5785498583aa39632beda3e7d76a8114d782f8ef3a9d207cd84c21"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.680253 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.680288 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.680299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"32a90e8c1cffa8c617f84a1f997c31ef3642da374d524220945443d43f1e80eb"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.682559 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.682591 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ac154dddb17fdd4144e1b7a52439efdaa9c5d66dd1ef362283526bd5bbf9e5bc"} Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.683196 4799 scope.go:117] "RemoveContainer" containerID="c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c" Sep 30 14:20:00 crc kubenswrapper[4799]: E0930 14:20:00.683316 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.684836 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.698838 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.712613 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.721218 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.736000 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.761635 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.775286 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.801055 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.824370 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.843161 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.862271 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.880772 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.895614 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.913190 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.930823 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.948867 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.979815 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:00 crc kubenswrapper[4799]: I0930 14:20:00.996304 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:00Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.014780 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.034103 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-vmvlc"] Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.034689 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: W0930 14:20:01.039027 4799 reflector.go:561] object-"openshift-image-registry"/"image-registry-certificates": failed to list *v1.ConfigMap: configmaps "image-registry-certificates" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-image-registry": no relationship found between node 'crc' and this object Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.039076 4799 reflector.go:158] "Unhandled Error" err="object-\"openshift-image-registry\"/\"image-registry-certificates\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-registry-certificates\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-image-registry\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.046322 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.050499 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.056594 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.070748 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.079903 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.089412 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.145531 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.145828 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.145925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5tgs\" (UniqueName: \"kubernetes.io/projected/06061f81-237c-4944-b7f2-0352d21ff125-kube-api-access-d5tgs\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.145999 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:20:03.145957178 +0000 UTC m=+25.229557595 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.146065 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.146135 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.146113 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06061f81-237c-4944-b7f2-0352d21ff125-host\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.146193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/06061f81-237c-4944-b7f2-0352d21ff125-serviceca\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.146230 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:03.146198705 +0000 UTC m=+25.229799212 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.160566 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.205938 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.241893 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.247419 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/06061f81-237c-4944-b7f2-0352d21ff125-serviceca\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.247475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5tgs\" (UniqueName: \"kubernetes.io/projected/06061f81-237c-4944-b7f2-0352d21ff125-kube-api-access-d5tgs\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.247520 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247700 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247749 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247774 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247728 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247819 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247828 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247832 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:03.247817948 +0000 UTC m=+25.331418375 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247857 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:03.247848199 +0000 UTC m=+25.331448626 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.247727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.247884 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.247911 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06061f81-237c-4944-b7f2-0352d21ff125-host\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.247947 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06061f81-237c-4944-b7f2-0352d21ff125-host\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.247980 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.248001 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:03.247993893 +0000 UTC m=+25.331594320 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.288076 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5tgs\" (UniqueName: \"kubernetes.io/projected/06061f81-237c-4944-b7f2-0352d21ff125-kube-api-access-d5tgs\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.297754 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.342864 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.386695 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.417260 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.458951 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.497763 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.502972 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.503043 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.503079 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:01 crc kubenswrapper[4799]: E0930 14:20:01.503243 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.535864 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.578788 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.627797 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.656123 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.686071 4799 generic.go:334] "Generic (PLEG): container finished" podID="cce844e9-25b0-49e7-9ca3-899b986d0f6f" containerID="94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1" exitCode=0 Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.686711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" event={"ID":"cce844e9-25b0-49e7-9ca3-899b986d0f6f","Type":"ContainerDied","Data":"94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1"} Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.688904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.688981 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.689004 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.689021 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.710415 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.737840 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.778102 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.821325 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.875405 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.904148 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.941926 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:01 crc kubenswrapper[4799]: I0930 14:20:01.980567 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:01Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.018270 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.057566 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.104372 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.150285 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.178524 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.229997 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: E0930 14:20:02.247720 4799 configmap.go:193] Couldn't get configMap openshift-image-registry/image-registry-certificates: failed to sync configmap cache: timed out waiting for the condition Sep 30 14:20:02 crc kubenswrapper[4799]: E0930 14:20:02.247848 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/06061f81-237c-4944-b7f2-0352d21ff125-serviceca podName:06061f81-237c-4944-b7f2-0352d21ff125 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:02.747822197 +0000 UTC m=+24.831422624 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serviceca" (UniqueName: "kubernetes.io/configmap/06061f81-237c-4944-b7f2-0352d21ff125-serviceca") pod "node-ca-vmvlc" (UID: "06061f81-237c-4944-b7f2-0352d21ff125") : failed to sync configmap cache: timed out waiting for the condition Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.257283 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.298297 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.309049 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.364794 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.400065 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.436843 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.487075 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.502229 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:02 crc kubenswrapper[4799]: E0930 14:20:02.502353 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.524758 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.555626 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.599115 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.693292 4799 generic.go:334] "Generic (PLEG): container finished" podID="cce844e9-25b0-49e7-9ca3-899b986d0f6f" containerID="0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b" exitCode=0 Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.693352 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" event={"ID":"cce844e9-25b0-49e7-9ca3-899b986d0f6f","Type":"ContainerDied","Data":"0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b"} Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.697093 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.697151 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.698298 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51"} Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.711269 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.729817 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.746791 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.763403 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/06061f81-237c-4944-b7f2-0352d21ff125-serviceca\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.764451 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.764624 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/06061f81-237c-4944-b7f2-0352d21ff125-serviceca\") pod \"node-ca-vmvlc\" (UID: \"06061f81-237c-4944-b7f2-0352d21ff125\") " pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.797777 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.841412 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.848442 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vmvlc" Sep 30 14:20:02 crc kubenswrapper[4799]: W0930 14:20:02.861149 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06061f81_237c_4944_b7f2_0352d21ff125.slice/crio-447d7605e89bfc59e29d937dea5a83c7e1bd18f32d3ae88f963f7ffef34b22bb WatchSource:0}: Error finding container 447d7605e89bfc59e29d937dea5a83c7e1bd18f32d3ae88f963f7ffef34b22bb: Status 404 returned error can't find the container with id 447d7605e89bfc59e29d937dea5a83c7e1bd18f32d3ae88f963f7ffef34b22bb Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.887834 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.936815 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.970611 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:02 crc kubenswrapper[4799]: I0930 14:20:02.999146 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:02Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.042551 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.078001 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.120226 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.160485 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.166912 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.167144 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:20:07.16710043 +0000 UTC m=+29.250700967 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.167241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.167450 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.167558 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:07.167530792 +0000 UTC m=+29.251131389 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.198589 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.241357 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.268352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.268395 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.268466 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268598 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268598 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268618 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268630 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268632 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268643 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268642 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268715 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:07.268698231 +0000 UTC m=+29.352298658 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268732 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:07.268727502 +0000 UTC m=+29.352327929 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.268778 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:07.268750753 +0000 UTC m=+29.352351350 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.277431 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.317446 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.364184 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.399819 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.438960 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.481447 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.502730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.502791 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.502930 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:03 crc kubenswrapper[4799]: E0930 14:20:03.503278 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.528858 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.560196 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.602579 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.643489 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.687995 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.704805 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vmvlc" event={"ID":"06061f81-237c-4944-b7f2-0352d21ff125","Type":"ContainerStarted","Data":"8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66"} Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.704880 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vmvlc" event={"ID":"06061f81-237c-4944-b7f2-0352d21ff125","Type":"ContainerStarted","Data":"447d7605e89bfc59e29d937dea5a83c7e1bd18f32d3ae88f963f7ffef34b22bb"} Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.709897 4799 generic.go:334] "Generic (PLEG): container finished" podID="cce844e9-25b0-49e7-9ca3-899b986d0f6f" containerID="61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f" exitCode=0 Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.710380 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" event={"ID":"cce844e9-25b0-49e7-9ca3-899b986d0f6f","Type":"ContainerDied","Data":"61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f"} Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.718215 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.763837 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.799431 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.842303 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.880092 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.919444 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.964946 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:03 crc kubenswrapper[4799]: I0930 14:20:03.998539 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:03Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.007843 4799 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.010092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.010150 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.010165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.010343 4799 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.054579 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.071293 4799 kubelet_node_status.go:115] "Node was previously registered" node="crc" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.071744 4799 kubelet_node_status.go:79] "Successfully registered node" node="crc" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.072956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.072996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.073012 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.073032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.073046 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: E0930 14:20:04.088589 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.092103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.092224 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.092281 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.092340 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.092407 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: E0930 14:20:04.110510 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.115364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.115403 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.115415 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.115431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.115442 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.122422 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: E0930 14:20:04.130157 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.135053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.135081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.135089 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.135102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.135113 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: E0930 14:20:04.151387 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.155145 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.155351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.155430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.155527 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.155616 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.168221 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: E0930 14:20:04.170389 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: E0930 14:20:04.170553 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.173329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.173361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.173370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.173384 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.173393 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.200348 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.247307 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.275611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.275645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.275666 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.275680 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.275689 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.278939 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.317909 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.359924 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.378347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.378387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.378395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.378410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.378419 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.402026 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.438362 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.480901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.481144 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.481223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.481293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.481391 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.502364 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:04 crc kubenswrapper[4799]: E0930 14:20:04.502639 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.583222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.583250 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.583259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.583274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.583283 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.684886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.684919 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.684928 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.684942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.684950 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.715318 4799 generic.go:334] "Generic (PLEG): container finished" podID="cce844e9-25b0-49e7-9ca3-899b986d0f6f" containerID="fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927" exitCode=0 Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.715388 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" event={"ID":"cce844e9-25b0-49e7-9ca3-899b986d0f6f","Type":"ContainerDied","Data":"fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.722241 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.733020 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.748826 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.760407 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.773635 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.787736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.787768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.787779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.787797 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.787808 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.791158 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.803577 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.817979 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.836304 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.848390 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.868120 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.882303 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.890715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.890752 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.890760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.890779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.890790 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.918201 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.958202 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.993417 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.993451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.993460 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.993477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.993488 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:04Z","lastTransitionTime":"2025-09-30T14:20:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:04 crc kubenswrapper[4799]: I0930 14:20:04.998155 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:04Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.037437 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.095958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.096331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.096484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.096607 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.096730 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.199111 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.199141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.199152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.199168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.199179 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.301970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.302157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.302213 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.302300 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.302356 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.404253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.404294 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.404304 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.404321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.404331 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.503837 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:05 crc kubenswrapper[4799]: E0930 14:20:05.504014 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.504471 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:05 crc kubenswrapper[4799]: E0930 14:20:05.504550 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.507223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.507265 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.507276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.507293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.507305 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.609536 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.609574 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.609585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.609603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.609615 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.712225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.712506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.712578 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.712668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.712734 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.727945 4799 generic.go:334] "Generic (PLEG): container finished" podID="cce844e9-25b0-49e7-9ca3-899b986d0f6f" containerID="2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947" exitCode=0 Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.727991 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" event={"ID":"cce844e9-25b0-49e7-9ca3-899b986d0f6f","Type":"ContainerDied","Data":"2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.741351 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.762249 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.788576 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.799025 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.817539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.817573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.817581 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.817598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.817607 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.818630 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.831272 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.840980 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.854405 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.867814 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.879874 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.894919 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.907210 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.918351 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.919687 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.919721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.919730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.919744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.919753 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:05Z","lastTransitionTime":"2025-09-30T14:20:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.931620 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:05 crc kubenswrapper[4799]: I0930 14:20:05.944173 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:05Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.022300 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.022339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.022349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.022364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.022373 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.124520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.124554 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.124562 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.124575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.124585 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.227292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.227331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.227345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.227360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.227372 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.330242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.330280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.330291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.330308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.330318 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.432726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.432761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.432769 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.432783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.432792 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.502821 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:06 crc kubenswrapper[4799]: E0930 14:20:06.502966 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.536021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.536284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.536373 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.536457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.536537 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.639281 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.639308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.639316 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.639330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.639340 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.734136 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" event={"ID":"cce844e9-25b0-49e7-9ca3-899b986d0f6f","Type":"ContainerStarted","Data":"89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.744364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.744805 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.744903 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.744969 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.745029 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.746070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.746764 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.746824 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.746840 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.752446 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.766494 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.776872 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.794774 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.804969 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.811494 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.812232 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.825828 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.839157 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.847869 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.847929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.847940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.847957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.847967 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.852133 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.866966 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.879530 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.890602 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.903457 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.913018 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.925256 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.937120 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.949722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.949773 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.949814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.949829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.949842 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:06Z","lastTransitionTime":"2025-09-30T14:20:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.949860 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.961497 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.976842 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:06 crc kubenswrapper[4799]: I0930 14:20:06.994149 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.005085 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.036583 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.052722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.052757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.052769 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.052785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.052795 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.067946 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.092232 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.121172 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.137248 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.155792 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.155846 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.155860 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.155884 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.155898 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.156643 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.173226 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.189809 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.202909 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.213856 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.214013 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.214088 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:20:15.214047485 +0000 UTC m=+37.297647912 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.214161 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.214225 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:15.21420925 +0000 UTC m=+37.297809677 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.217783 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:07Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.258645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.258712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.258722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.258743 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.258755 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.315251 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.315302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.315323 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315415 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315467 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:15.315452802 +0000 UTC m=+37.399053219 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315476 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315506 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315518 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315569 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:15.315553875 +0000 UTC m=+37.399154302 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315631 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315640 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315665 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.315688 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:15.315679438 +0000 UTC m=+37.399279865 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.363262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.363322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.363331 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.363350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.363360 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.466029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.466068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.466080 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.466095 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.466128 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.502902 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.503045 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.504780 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:07 crc kubenswrapper[4799]: E0930 14:20:07.504961 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.568784 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.568822 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.568830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.568844 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.568853 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.672055 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.672085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.672094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.672110 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.672120 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.775923 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.775991 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.776006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.776029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.776046 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.878460 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.878505 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.878515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.878539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.878555 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.981733 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.981772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.981782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.981801 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:07 crc kubenswrapper[4799]: I0930 14:20:07.981812 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:07Z","lastTransitionTime":"2025-09-30T14:20:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.084851 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.084942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.084971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.085001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.085022 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.187256 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.187474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.187691 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.187789 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.187851 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.289911 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.289956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.289965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.289983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.289993 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.392498 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.392725 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.392821 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.392884 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.392956 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.494712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.494750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.494762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.494778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.494789 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.502399 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:08 crc kubenswrapper[4799]: E0930 14:20:08.502697 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.514310 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.529466 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.540544 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.555120 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.568590 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.584388 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.596710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.596751 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.596760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.596776 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.596785 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.596905 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.608367 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.622611 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.639016 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.652437 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.669054 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.690124 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.698919 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.698992 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.699004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.699018 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.699029 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.708154 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.728315 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.801173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.801205 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.801231 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.801246 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.801254 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.903813 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.903853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.903863 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.903877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:08 crc kubenswrapper[4799]: I0930 14:20:08.903888 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:08Z","lastTransitionTime":"2025-09-30T14:20:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.005842 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.005878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.005889 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.005902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.005912 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.108568 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.108608 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.108616 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.108630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.108640 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.210165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.210195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.210206 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.210222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.210233 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.312616 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.312645 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.312663 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.312675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.312684 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.415594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.415640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.415659 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.415698 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.415707 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.502904 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.502949 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:09 crc kubenswrapper[4799]: E0930 14:20:09.503020 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:09 crc kubenswrapper[4799]: E0930 14:20:09.503104 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.518233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.518280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.518292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.518308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.518319 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.621147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.621194 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.621203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.621221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.621230 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.724033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.724076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.724086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.724104 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.724121 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.756800 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/0.log" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.759583 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b" exitCode=1 Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.759622 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.762568 4799 scope.go:117] "RemoveContainer" containerID="b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.781884 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.795020 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.810148 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.825946 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.827049 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.827085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.827094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.827109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.827118 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.839040 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.850801 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.864752 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.887951 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:09Z\\\",\\\"message\\\":\\\"nition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0930 14:20:09.345074 6023 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345260 6023 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 14:20:09.345337 6023 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:09.345677 6023 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345927 6023 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345958 6023 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 14:20:09.345972 6023 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 14:20:09.345986 6023 factory.go:656] Stopping watch factory\\\\nI0930 14:20:09.346014 6023 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 14:20:09.346029 6023 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.899387 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.916864 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.927184 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.929312 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.929333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.929342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.929358 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.929367 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:09Z","lastTransitionTime":"2025-09-30T14:20:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.936855 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.953108 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.966234 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:09 crc kubenswrapper[4799]: I0930 14:20:09.981199 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:09Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.031835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.031874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.031884 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.031899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.031908 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.133851 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.133891 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.133901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.133918 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.133927 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.236014 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.236054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.236063 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.236081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.236091 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.337785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.337981 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.338073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.338139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.338195 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.440616 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.440850 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.440942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.441135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.441223 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.502436 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:10 crc kubenswrapper[4799]: E0930 14:20:10.502843 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.543617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.543675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.543695 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.543713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.543725 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.646103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.646136 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.646146 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.646159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.646168 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.748408 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.748441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.748450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.748466 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.748481 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.765322 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/1.log" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.765939 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/0.log" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.769009 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543" exitCode=1 Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.769060 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.769107 4799 scope.go:117] "RemoveContainer" containerID="b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.769723 4799 scope.go:117] "RemoveContainer" containerID="0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543" Sep 30 14:20:10 crc kubenswrapper[4799]: E0930 14:20:10.769879 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.789294 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.802928 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.818303 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.830394 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.845294 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.850569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.850600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.850609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.850624 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.850635 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.856748 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.868221 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.881630 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.892942 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.903524 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.921199 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.933254 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.943566 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.952457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.952596 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.952679 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.952742 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.952795 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:10Z","lastTransitionTime":"2025-09-30T14:20:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.959164 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:10 crc kubenswrapper[4799]: I0930 14:20:10.980730 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:09Z\\\",\\\"message\\\":\\\"nition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0930 14:20:09.345074 6023 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345260 6023 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 14:20:09.345337 6023 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:09.345677 6023 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345927 6023 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345958 6023 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 14:20:09.345972 6023 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 14:20:09.345986 6023 factory.go:656] Stopping watch factory\\\\nI0930 14:20:09.346014 6023 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 14:20:09.346029 6023 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:10Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.055633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.055688 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.055700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.055719 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.055730 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.069585 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8"] Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.070008 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.073926 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.074662 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.085557 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.097502 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.111669 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.133223 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:09Z\\\",\\\"message\\\":\\\"nition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0930 14:20:09.345074 6023 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345260 6023 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 14:20:09.345337 6023 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:09.345677 6023 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345927 6023 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345958 6023 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 14:20:09.345972 6023 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 14:20:09.345986 6023 factory.go:656] Stopping watch factory\\\\nI0930 14:20:09.346014 6023 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 14:20:09.346029 6023 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.142725 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.157890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.157921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.157929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.157943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.157952 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.164298 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.180225 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.193166 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.204021 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.218505 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.231923 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.244114 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.250999 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.251030 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgr2b\" (UniqueName: \"kubernetes.io/projected/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-kube-api-access-wgr2b\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.251070 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.251105 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.256828 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.260209 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.260342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.260423 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.260510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.260594 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.269199 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.281119 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.294440 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:11Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.352132 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.352383 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgr2b\" (UniqueName: \"kubernetes.io/projected/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-kube-api-access-wgr2b\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.352522 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.352595 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.353404 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.353379 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.366583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.367507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.367541 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.367577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.367587 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.367780 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.372106 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgr2b\" (UniqueName: \"kubernetes.io/projected/d16ad3a5-b9c0-42a6-aff2-11b58fb566fc-kube-api-access-wgr2b\") pod \"ovnkube-control-plane-749d76644c-7vhf8\" (UID: \"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.383448 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.470015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.470063 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.470072 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.470089 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.470101 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.502773 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.502838 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:11 crc kubenswrapper[4799]: E0930 14:20:11.502901 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:11 crc kubenswrapper[4799]: E0930 14:20:11.503125 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.573031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.573062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.573070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.573084 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.573092 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.677293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.677330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.677342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.677357 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.677368 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.774599 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/1.log" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.778712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.778775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.778787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.778824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.778836 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.780739 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" event={"ID":"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc","Type":"ContainerStarted","Data":"4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.780784 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" event={"ID":"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc","Type":"ContainerStarted","Data":"7ebbfa0b0b2b0c321cd55adffb3238b9407316e866010f9de2630a9ddc662bc3"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.880730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.880770 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.880779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.880796 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.880806 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.983076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.983114 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.983126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.983144 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:11 crc kubenswrapper[4799]: I0930 14:20:11.983155 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:11Z","lastTransitionTime":"2025-09-30T14:20:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.085754 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.085795 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.085807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.085826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.085837 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.188247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.188297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.188310 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.188329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.188341 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.290625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.290735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.290746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.290763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.290774 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.393514 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.393555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.393565 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.393581 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.393592 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.495710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.495945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.496035 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.496182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.496287 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.503884 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:12 crc kubenswrapper[4799]: E0930 14:20:12.503981 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.538512 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hghbr"] Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.539251 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:12 crc kubenswrapper[4799]: E0930 14:20:12.539396 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.551355 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.564766 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.575721 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.588251 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.598023 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.598049 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.598057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.598071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.598081 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.601603 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.615414 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.638792 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:09Z\\\",\\\"message\\\":\\\"nition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0930 14:20:09.345074 6023 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345260 6023 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 14:20:09.345337 6023 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:09.345677 6023 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345927 6023 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345958 6023 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 14:20:09.345972 6023 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 14:20:09.345986 6023 factory.go:656] Stopping watch factory\\\\nI0930 14:20:09.346014 6023 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 14:20:09.346029 6023 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.648829 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.665597 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.666052 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.666191 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bkzz\" (UniqueName: \"kubernetes.io/projected/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-kube-api-access-8bkzz\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.677177 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.686548 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.697730 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.700541 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.700571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.700580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.700592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.700601 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.709236 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.720100 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.731250 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.743746 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.756633 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.767054 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bkzz\" (UniqueName: \"kubernetes.io/projected/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-kube-api-access-8bkzz\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.767118 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:12 crc kubenswrapper[4799]: E0930 14:20:12.767196 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:12 crc kubenswrapper[4799]: E0930 14:20:12.767236 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs podName:15ddd33f-0b7f-4cb5-b09a-03c211c759c8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:13.2672234 +0000 UTC m=+35.350823827 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs") pod "network-metrics-daemon-hghbr" (UID: "15ddd33f-0b7f-4cb5-b09a-03c211c759c8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.785096 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" event={"ID":"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc","Type":"ContainerStarted","Data":"082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.786577 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bkzz\" (UniqueName: \"kubernetes.io/projected/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-kube-api-access-8bkzz\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.802859 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.803094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.803164 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.803223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.803282 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.803934 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.815319 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.824385 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.835876 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.851617 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:09Z\\\",\\\"message\\\":\\\"nition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0930 14:20:09.345074 6023 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345260 6023 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 14:20:09.345337 6023 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:09.345677 6023 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345927 6023 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345958 6023 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 14:20:09.345972 6023 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 14:20:09.345986 6023 factory.go:656] Stopping watch factory\\\\nI0930 14:20:09.346014 6023 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 14:20:09.346029 6023 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.860752 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.881639 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.894165 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.905940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.906017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.906029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.906044 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.906056 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:12Z","lastTransitionTime":"2025-09-30T14:20:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.907173 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.916519 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.932364 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.944675 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.957136 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.967621 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.982395 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:12 crc kubenswrapper[4799]: I0930 14:20:12.997294 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:12Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.009806 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.009868 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.009877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.009894 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.009904 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.012107 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.112025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.112071 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.112083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.112101 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.112111 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.215285 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.215358 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.215377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.215407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.215426 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.274536 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:13 crc kubenswrapper[4799]: E0930 14:20:13.274771 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:13 crc kubenswrapper[4799]: E0930 14:20:13.274848 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs podName:15ddd33f-0b7f-4cb5-b09a-03c211c759c8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:14.274825461 +0000 UTC m=+36.358425898 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs") pod "network-metrics-daemon-hghbr" (UID: "15ddd33f-0b7f-4cb5-b09a-03c211c759c8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.317880 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.318270 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.318410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.318502 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.318598 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.422424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.422495 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.422518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.422547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.422568 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.503361 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.503369 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:13 crc kubenswrapper[4799]: E0930 14:20:13.503745 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:13 crc kubenswrapper[4799]: E0930 14:20:13.503773 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.504615 4799 scope.go:117] "RemoveContainer" containerID="c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.525997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.526054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.526067 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.526091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.526105 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.628871 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.629390 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.629400 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.629418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.629429 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.732908 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.732945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.732954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.732971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.732980 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.790344 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.792354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.792969 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.805724 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.821506 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.836139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.836181 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.836192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.836210 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.836222 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.839962 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:09Z\\\",\\\"message\\\":\\\"nition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0930 14:20:09.345074 6023 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345260 6023 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 14:20:09.345337 6023 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:09.345677 6023 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345927 6023 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345958 6023 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 14:20:09.345972 6023 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 14:20:09.345986 6023 factory.go:656] Stopping watch factory\\\\nI0930 14:20:09.346014 6023 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 14:20:09.346029 6023 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.851499 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.870381 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.882736 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.892776 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.904024 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.915339 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.926557 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.936758 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.938395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.938593 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.938699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.938781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.938868 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:13Z","lastTransitionTime":"2025-09-30T14:20:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.951524 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.964078 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.977490 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:13 crc kubenswrapper[4799]: I0930 14:20:13.989462 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:13Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.008589 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:14Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.019541 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:14Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.041418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.041451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.041471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.041488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.041498 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.145110 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.145148 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.145158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.145173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.145184 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.247153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.247184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.247192 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.247207 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.247218 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.285962 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.286126 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.286344 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs podName:15ddd33f-0b7f-4cb5-b09a-03c211c759c8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:16.286319942 +0000 UTC m=+38.369920379 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs") pod "network-metrics-daemon-hghbr" (UID: "15ddd33f-0b7f-4cb5-b09a-03c211c759c8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.349351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.349385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.349395 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.349409 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.349418 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.396774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.397010 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.397075 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.397206 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.397287 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.412202 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:14Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.415759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.416035 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.416113 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.416201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.416299 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.430924 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:14Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.436163 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.436202 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.436211 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.436226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.436237 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.449921 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:14Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.455012 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.455078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.455092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.455110 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.455120 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.470211 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:14Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.475426 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.475467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.475478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.475497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.475507 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.493861 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:14Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.493993 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.496661 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.496710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.496720 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.496739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.496749 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.503040 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.503050 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.503210 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:14 crc kubenswrapper[4799]: E0930 14:20:14.503342 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.599589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.599637 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.599673 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.599699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.599717 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.702394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.702711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.702816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.702882 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.702938 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.805697 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.805738 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.805747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.805762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.805773 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.908673 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.908719 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.908730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.908746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:14 crc kubenswrapper[4799]: I0930 14:20:14.908764 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:14Z","lastTransitionTime":"2025-09-30T14:20:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.011248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.011283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.011292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.011309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.011318 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.113081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.113124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.113136 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.113153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.113162 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.215813 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.215851 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.215860 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.215874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.215888 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.297269 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.297444 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.297505 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:20:31.297487895 +0000 UTC m=+53.381088322 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.297601 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.297721 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:31.297700181 +0000 UTC m=+53.381300678 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.317826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.317854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.317865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.317882 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.317894 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.398281 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.398707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.398899 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.398637 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.399160 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.399269 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.398867 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.399392 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.399410 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.399465 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:31.399447458 +0000 UTC m=+53.483047895 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.399009 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.399733 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:31.399719716 +0000 UTC m=+53.483320143 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.399933 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:31.399920521 +0000 UTC m=+53.483520958 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.420167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.420208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.420222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.420242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.420257 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.502804 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.502946 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.503263 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:15 crc kubenswrapper[4799]: E0930 14:20:15.503313 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.522489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.522520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.522529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.522542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.522551 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.625974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.626091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.626155 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.626188 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.626207 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.728944 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.729004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.729016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.729033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.729045 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.831717 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.831749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.831758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.831771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.831780 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.934254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.934291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.934303 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.934318 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:15 crc kubenswrapper[4799]: I0930 14:20:15.934329 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:15Z","lastTransitionTime":"2025-09-30T14:20:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.036311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.036597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.036726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.036830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.036890 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.139621 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.139688 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.139700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.139713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.139722 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.242458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.242513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.242536 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.242569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.242586 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.307299 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:16 crc kubenswrapper[4799]: E0930 14:20:16.307459 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:16 crc kubenswrapper[4799]: E0930 14:20:16.307505 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs podName:15ddd33f-0b7f-4cb5-b09a-03c211c759c8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:20.307491884 +0000 UTC m=+42.391092311 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs") pod "network-metrics-daemon-hghbr" (UID: "15ddd33f-0b7f-4cb5-b09a-03c211c759c8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.344814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.344843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.344852 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.344865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.344875 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.447196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.447229 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.447239 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.447251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.447261 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.503779 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:16 crc kubenswrapper[4799]: E0930 14:20:16.504114 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.503919 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:16 crc kubenswrapper[4799]: E0930 14:20:16.504392 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.552980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.553019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.553029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.553043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.553052 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.655005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.655091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.655102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.655125 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.655143 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.757202 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.757241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.757252 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.757267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.757277 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.859422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.859477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.859498 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.859517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.859531 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.962607 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.962662 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.962672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.962686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:16 crc kubenswrapper[4799]: I0930 14:20:16.962697 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:16Z","lastTransitionTime":"2025-09-30T14:20:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.064422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.064464 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.064476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.064494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.064506 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.166608 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.166694 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.166726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.166745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.166756 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.269958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.270003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.270015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.270031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.270043 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.372721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.372761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.372770 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.372784 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.372793 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.474973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.475027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.475038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.475056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.475072 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.502292 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:17 crc kubenswrapper[4799]: E0930 14:20:17.502426 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.502311 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:17 crc kubenswrapper[4799]: E0930 14:20:17.502499 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.578516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.578556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.578565 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.578583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.578606 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.680734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.680762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.680770 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.680786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.680795 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.783484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.783532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.783547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.783566 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.783580 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.886108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.886158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.886171 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.886189 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.886200 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.988971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.989009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.989017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.989029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:17 crc kubenswrapper[4799]: I0930 14:20:17.989039 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:17Z","lastTransitionTime":"2025-09-30T14:20:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.091545 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.091816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.091887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.091964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.092038 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.194758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.195047 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.195111 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.195200 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.195266 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.297398 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.297443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.297454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.297467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.297477 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.399277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.399319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.399330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.399344 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.399354 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.501936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.502238 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.502244 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.502399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.502499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.502584 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.502708 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:18 crc kubenswrapper[4799]: E0930 14:20:18.502773 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:18 crc kubenswrapper[4799]: E0930 14:20:18.502325 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.518991 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.529831 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.543924 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.561890 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b09be3d2ae0a9a5a8943aaf4ea3a495a5366f5d98872e30450790b2fa6ed9f1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:09Z\\\",\\\"message\\\":\\\"nition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0930 14:20:09.345074 6023 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345260 6023 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0930 14:20:09.345337 6023 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:09.345677 6023 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345927 6023 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:09.345958 6023 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0930 14:20:09.345972 6023 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0930 14:20:09.345986 6023 factory.go:656] Stopping watch factory\\\\nI0930 14:20:09.346014 6023 handler.go:208] Removed *v1.Node event handler 2\\\\nI0930 14:20:09.346029 6023 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.575218 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.601224 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.605551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.605740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.605764 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.605783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.605797 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.614774 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.629427 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.644189 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.659284 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.672173 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.685012 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.695563 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.708731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.708785 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.708795 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.708819 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.708867 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.710011 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.725144 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.740743 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.760434 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:18Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.812025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.812299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.812428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.812551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.812714 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.915718 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.915753 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.915762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.915776 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:18 crc kubenswrapper[4799]: I0930 14:20:18.915786 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:18Z","lastTransitionTime":"2025-09-30T14:20:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.019383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.019446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.019463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.019487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.019503 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.122687 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.123080 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.123178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.123292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.123394 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.226405 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.226683 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.226980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.227184 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.227374 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.329970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.330009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.330020 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.330037 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.330048 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.432614 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.432643 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.432667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.432680 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.432689 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.503072 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:19 crc kubenswrapper[4799]: E0930 14:20:19.503207 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.503091 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:19 crc kubenswrapper[4799]: E0930 14:20:19.503287 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.535366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.535610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.535730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.535821 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.535928 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.638443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.638488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.638499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.638513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.638522 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.741034 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.741081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.741094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.741114 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.741167 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.843424 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.843474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.843486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.843505 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.843518 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.945949 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.946012 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.946024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.946041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:19 crc kubenswrapper[4799]: I0930 14:20:19.946084 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:19Z","lastTransitionTime":"2025-09-30T14:20:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.048258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.048309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.048322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.048346 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.048356 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.150919 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.150979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.150990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.151006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.151017 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.253663 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.253699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.253708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.253726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.253737 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.346120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:20 crc kubenswrapper[4799]: E0930 14:20:20.346293 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:20 crc kubenswrapper[4799]: E0930 14:20:20.346539 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs podName:15ddd33f-0b7f-4cb5-b09a-03c211c759c8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:28.346522219 +0000 UTC m=+50.430122646 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs") pod "network-metrics-daemon-hghbr" (UID: "15ddd33f-0b7f-4cb5-b09a-03c211c759c8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.355984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.356027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.356037 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.356057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.356068 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.458468 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.458523 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.458536 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.458574 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.458588 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.503080 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.503154 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:20 crc kubenswrapper[4799]: E0930 14:20:20.503239 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:20 crc kubenswrapper[4799]: E0930 14:20:20.503384 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.560985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.561027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.561037 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.561054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.561065 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.664880 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.664927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.664936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.664953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.664964 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.768039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.768092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.768106 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.768126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.768138 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.869833 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.869874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.869885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.869901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.869915 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.973124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.973183 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.973195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.973215 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:20 crc kubenswrapper[4799]: I0930 14:20:20.973227 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:20Z","lastTransitionTime":"2025-09-30T14:20:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.075951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.075992 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.076005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.076021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.076032 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.178223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.178271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.178283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.178301 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.178317 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.281195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.281242 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.281253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.281270 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.281281 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.384121 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.384148 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.384170 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.384183 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.384193 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.491658 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.491734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.491748 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.491766 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.491783 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.502158 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:21 crc kubenswrapper[4799]: E0930 14:20:21.502267 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.502158 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:21 crc kubenswrapper[4799]: E0930 14:20:21.502545 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.593881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.593915 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.593924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.593939 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.593949 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.696061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.696112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.696122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.696136 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.696147 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.798006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.798032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.798042 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.798073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.798081 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.900792 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.900830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.900841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.900855 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:21 crc kubenswrapper[4799]: I0930 14:20:21.900865 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:21Z","lastTransitionTime":"2025-09-30T14:20:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.003251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.003524 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.003615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.003710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.003907 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.106617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.107250 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.107388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.107472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.107554 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.210252 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.210290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.210302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.210317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.210328 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.312781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.312824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.312835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.312852 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.312864 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.415293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.415321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.415329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.415343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.415352 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.502906 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:22 crc kubenswrapper[4799]: E0930 14:20:22.503046 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.503306 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:22 crc kubenswrapper[4799]: E0930 14:20:22.503516 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.518378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.518407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.518415 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.518428 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.518440 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.621122 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.621439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.621520 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.621608 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.621702 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.724105 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.724359 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.724455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.724577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.724670 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.827530 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.827573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.827585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.827612 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.827626 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.930124 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.930167 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.930179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.930196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:22 crc kubenswrapper[4799]: I0930 14:20:22.930208 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:22Z","lastTransitionTime":"2025-09-30T14:20:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.032502 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.032568 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.032578 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.032596 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.032605 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.135084 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.135121 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.135132 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.135147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.135158 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.237535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.237580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.237593 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.237608 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.237618 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.340010 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.340076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.340088 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.340107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.340122 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.442334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.442377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.442386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.442399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.442408 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.502456 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.502531 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:23 crc kubenswrapper[4799]: E0930 14:20:23.502598 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:23 crc kubenswrapper[4799]: E0930 14:20:23.502748 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.504238 4799 scope.go:117] "RemoveContainer" containerID="0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.519845 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.540359 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.545032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.545282 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.545367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.545449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.545538 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.552638 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.562959 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.578004 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.592405 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.605625 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.616936 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.637927 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.648486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.648507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.648515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.648528 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.648536 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.653934 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.667364 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.681379 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.702306 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.714297 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.724978 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.737741 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.749488 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.750326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.750358 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.750369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.750386 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.750397 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.824169 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/1.log" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.827037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.827554 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.841701 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.852948 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.853066 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.853109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.853121 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.853139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.853150 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.862453 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.875269 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.887844 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.898978 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.920184 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.935018 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.951835 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.955680 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.955904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.956015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.956129 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.956233 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:23Z","lastTransitionTime":"2025-09-30T14:20:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.968500 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:23 crc kubenswrapper[4799]: I0930 14:20:23.998380 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:23Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.017905 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.037552 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.050381 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.058241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.058272 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.058280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.058293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.058303 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.064592 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.075853 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.091245 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.160984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.161020 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.161031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.161050 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.161060 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.263325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.263367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.263377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.263392 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.263404 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.365439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.365482 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.365496 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.365519 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.365533 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.468456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.468507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.468517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.468536 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.468547 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.502478 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.502496 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.502728 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.502782 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.571277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.571326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.571339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.571394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.571411 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.673351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.673389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.673400 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.673413 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.673423 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.775904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.775946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.775958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.775976 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.775988 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.781600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.781639 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.781667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.781684 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.781695 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.795506 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.798765 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.798794 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.798803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.798819 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.798828 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.811487 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.815329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.815362 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.815373 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.815390 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.815403 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.827699 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.831682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.831720 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.831731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.831750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.831763 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.833364 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/2.log" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.833923 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/1.log" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.836072 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32" exitCode=1 Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.836105 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.836137 4799 scope.go:117] "RemoveContainer" containerID="0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.836790 4799 scope.go:117] "RemoveContainer" containerID="af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32" Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.836924 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.845300 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.848525 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.852795 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.852830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.852842 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.852859 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.852869 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.862815 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.864161 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: E0930 14:20:24.864276 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.874025 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.877762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.877781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.877790 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.877804 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.877813 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.885591 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.898195 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.910227 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.921115 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.931764 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.945340 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.958830 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.969440 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.979951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.979989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.980000 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.980017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.980027 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:24Z","lastTransitionTime":"2025-09-30T14:20:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:24 crc kubenswrapper[4799]: I0930 14:20:24.984810 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.001175 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:24Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.011896 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.030259 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.040732 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.050245 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.082226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.082278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.082287 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.082302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.082311 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.184874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.184912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.184920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.184933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.184941 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.218562 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.234297 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.245981 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.261007 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.270321 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.281968 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.287234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.287459 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.287578 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.287726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.287858 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.294352 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.305872 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.315271 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.326979 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.339380 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.351698 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.364179 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.383493 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.390027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.390159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.390253 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.390324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.390411 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.397283 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.408455 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.423611 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.442171 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0137948a95082bbf7bcb7fd3f0075934203b54e32313025a5a3aec185fd1b543\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:10Z\\\",\\\"message\\\":\\\"r e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006bcc8f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.493699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.493951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.494013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.494079 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.494136 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.502248 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.502244 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:25 crc kubenswrapper[4799]: E0930 14:20:25.502723 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:25 crc kubenswrapper[4799]: E0930 14:20:25.502932 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.596782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.596826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.596836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.596850 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.596860 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.699259 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.699298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.699307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.699321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.699329 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.802377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.802921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.803082 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.803290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.803484 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.841323 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/2.log" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.845031 4799 scope.go:117] "RemoveContainer" containerID="af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32" Sep 30 14:20:25 crc kubenswrapper[4799]: E0930 14:20:25.845245 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.861389 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.875732 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.890138 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.903769 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.905889 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.905926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.905938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.905957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.905982 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:25Z","lastTransitionTime":"2025-09-30T14:20:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.920729 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.934137 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.949572 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.970433 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.983352 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:25 crc kubenswrapper[4799]: I0930 14:20:25.995203 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:25Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.009140 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.009368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.009443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.009517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.009589 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.012832 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:26Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.031178 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:26Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.045501 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:26Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.063223 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:26Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.077771 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:26Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.090846 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:26Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.102122 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:26Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.111907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.111949 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.111959 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.111973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.111981 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.214474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.214538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.214551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.214568 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.214580 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.316936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.316968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.316976 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.316989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.316998 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.419467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.419508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.419519 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.419537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.419549 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.503009 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.503045 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:26 crc kubenswrapper[4799]: E0930 14:20:26.503159 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:26 crc kubenswrapper[4799]: E0930 14:20:26.503254 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.521746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.522001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.522065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.522169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.522289 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.624509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.624806 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.624898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.624985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.625067 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.726852 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.726888 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.726899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.726916 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.726927 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.829636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.829709 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.829720 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.829738 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.829748 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.931355 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.931382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.931389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.931402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:26 crc kubenswrapper[4799]: I0930 14:20:26.931411 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:26Z","lastTransitionTime":"2025-09-30T14:20:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.033065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.033097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.033107 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.033123 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.033134 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.134940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.135008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.135019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.135057 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.135071 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.237750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.238321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.238429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.238534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.238673 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.340401 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.340477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.340490 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.340511 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.340526 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.443378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.443422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.443433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.443447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.443456 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.502769 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.502777 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:27 crc kubenswrapper[4799]: E0930 14:20:27.502917 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:27 crc kubenswrapper[4799]: E0930 14:20:27.503032 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.545797 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.545839 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.545848 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.545865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.545878 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.648843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.648900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.648910 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.648926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.648936 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.751500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.751729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.751837 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.751902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.751966 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.854573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.854612 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.854624 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.854640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.854668 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.957425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.957463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.957495 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.957512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:27 crc kubenswrapper[4799]: I0930 14:20:27.957520 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:27Z","lastTransitionTime":"2025-09-30T14:20:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.059433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.059641 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.059733 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.059820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.059962 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.162332 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.162560 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.162662 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.162745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.162858 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.265740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.265781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.265793 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.265809 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.265820 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.368609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.368666 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.368677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.368690 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.368700 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.428378 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:28 crc kubenswrapper[4799]: E0930 14:20:28.428510 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:28 crc kubenswrapper[4799]: E0930 14:20:28.428576 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs podName:15ddd33f-0b7f-4cb5-b09a-03c211c759c8 nodeName:}" failed. No retries permitted until 2025-09-30 14:20:44.428557026 +0000 UTC m=+66.512157453 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs") pod "network-metrics-daemon-hghbr" (UID: "15ddd33f-0b7f-4cb5-b09a-03c211c759c8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.471303 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.471512 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.471579 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.471640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.471732 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.502302 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.502387 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:28 crc kubenswrapper[4799]: E0930 14:20:28.502419 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:28 crc kubenswrapper[4799]: E0930 14:20:28.502571 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.524831 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.537577 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.549379 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.563279 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.573462 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.573704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.573774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.573894 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.573954 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.584758 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.594724 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.605195 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.615912 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.629377 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.638788 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.650327 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.662293 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.676336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.676387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.676399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.676417 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.676428 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.678228 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.689498 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.701470 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.714946 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.726725 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:28Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.780820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.780875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.780894 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.780914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.780929 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.883925 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.883968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.883979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.883996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.884007 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.986817 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.986848 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.986858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.986876 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:28 crc kubenswrapper[4799]: I0930 14:20:28.986886 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:28Z","lastTransitionTime":"2025-09-30T14:20:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.089206 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.089240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.089249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.089263 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.089275 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.191742 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.191779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.191789 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.191804 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.191815 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.294763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.294813 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.294824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.294841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.294850 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.396823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.396848 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.396856 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.396870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.396878 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.498960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.498994 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.499005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.499026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.499034 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.502222 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:29 crc kubenswrapper[4799]: E0930 14:20:29.502345 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.502644 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:29 crc kubenswrapper[4799]: E0930 14:20:29.502727 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.602086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.602161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.602195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.602214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.602227 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.705583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.705665 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.705677 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.705701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.705720 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.809828 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.810075 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.810188 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.810297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.810387 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.914180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.914298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.914314 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.914338 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:29 crc kubenswrapper[4799]: I0930 14:20:29.914379 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:29Z","lastTransitionTime":"2025-09-30T14:20:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.016779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.016854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.016865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.016885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.016900 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.119459 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.119563 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.119579 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.119620 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.119637 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.222765 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.222807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.222818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.222835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.222846 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.325580 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.325635 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.325662 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.325698 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.325714 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.429507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.429590 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.429602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.429630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.429678 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.502726 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.502737 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:30 crc kubenswrapper[4799]: E0930 14:20:30.503018 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:30 crc kubenswrapper[4799]: E0930 14:20:30.503084 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.532419 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.532657 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.532744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.532810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.532883 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.635403 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.635777 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.635789 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.635807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.635819 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.738248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.738289 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.738297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.738311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.738321 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.840445 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.840473 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.840480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.840493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.840500 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.944174 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.944227 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.944236 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.944250 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:30 crc kubenswrapper[4799]: I0930 14:20:30.944259 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:30Z","lastTransitionTime":"2025-09-30T14:20:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.045968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.046001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.046010 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.046024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.046033 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.148563 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.148610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.148625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.148643 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.148689 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.251948 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.251993 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.252002 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.252018 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.252028 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.354625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.354702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.354715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.354735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.354747 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.360769 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.360915 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.361043 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.361094 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:21:03.36108129 +0000 UTC m=+85.444681717 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.361239 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:21:03.361231935 +0000 UTC m=+85.444832362 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.457189 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.457216 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.457225 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.457238 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.457251 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.462173 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.462207 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.462226 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462291 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462329 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:21:03.462316552 +0000 UTC m=+85.545916979 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462471 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462488 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462496 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462519 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 14:21:03.462511807 +0000 UTC m=+85.546112234 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462844 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462862 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462870 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.462893 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 14:21:03.462884948 +0000 UTC m=+85.546485375 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.502818 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.502914 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.502970 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:31 crc kubenswrapper[4799]: E0930 14:20:31.503053 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.560026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.560072 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.560081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.560099 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.560109 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.663980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.664038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.664054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.664079 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.664098 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.766468 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.766533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.766544 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.766558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.766567 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.868290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.868564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.868632 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.868740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.868801 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.971672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.971711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.971726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.971745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:31 crc kubenswrapper[4799]: I0930 14:20:31.971757 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:31Z","lastTransitionTime":"2025-09-30T14:20:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.074148 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.074180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.074189 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.074203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.074213 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.176807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.176845 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.176855 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.176870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.176880 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.278924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.278957 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.278965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.278981 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.278991 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.381688 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.381735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.381747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.381764 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.381774 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.483286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.483314 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.483323 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.483336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.483345 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.502797 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:32 crc kubenswrapper[4799]: E0930 14:20:32.502893 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.503297 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:32 crc kubenswrapper[4799]: E0930 14:20:32.503357 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.585501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.585536 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.585545 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.585558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.585568 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.688059 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.688090 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.688098 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.688112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.688121 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.753973 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.763929 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.768140 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.779951 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.791436 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.791975 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.792022 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.792075 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.792097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.792109 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.802436 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.816558 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.835117 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.845218 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.864334 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.880695 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.891345 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.893958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.894005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.894014 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.894031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.894040 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:32Z","lastTransitionTime":"2025-09-30T14:20:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.905594 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.917544 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.932026 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.941509 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.952678 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.963772 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:32 crc kubenswrapper[4799]: I0930 14:20:32.973938 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:32Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.000952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.001000 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.001018 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.001035 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.001047 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.102924 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.102960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.102971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.102986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.102995 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.206151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.206191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.206201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.206218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.206227 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.308115 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.308152 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.308160 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.308174 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.308183 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.410504 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.410546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.410555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.410569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.410580 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.502502 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.502540 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:33 crc kubenswrapper[4799]: E0930 14:20:33.502625 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:33 crc kubenswrapper[4799]: E0930 14:20:33.502754 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.512551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.512599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.512610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.512627 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.512643 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.614820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.614854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.614864 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.614877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.614886 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.717322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.717364 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.717375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.717392 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.717403 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.819748 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.819805 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.819821 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.819842 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.819857 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.922526 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.922594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.922609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.922626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:33 crc kubenswrapper[4799]: I0930 14:20:33.922638 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:33Z","lastTransitionTime":"2025-09-30T14:20:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.025208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.025247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.025261 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.025283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.025295 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.128436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.128499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.128509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.128525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.128535 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.230733 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.230768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.230778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.230791 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.230800 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.333510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.333557 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.333570 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.333588 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.333603 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.450862 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.450897 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.450906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.450920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.450929 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.503873 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.503912 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:34 crc kubenswrapper[4799]: E0930 14:20:34.503991 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:34 crc kubenswrapper[4799]: E0930 14:20:34.504324 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.552876 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.552914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.552925 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.552951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.552992 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.655508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.655548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.655557 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.655570 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.655581 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.757844 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.757878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.757890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.757908 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.757918 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.860256 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.860291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.860306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.860322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.860332 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.962816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.962850 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.962858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.962871 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:34 crc kubenswrapper[4799]: I0930 14:20:34.962879 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:34Z","lastTransitionTime":"2025-09-30T14:20:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.065346 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.065383 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.065403 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.065431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.065445 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.167840 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.167875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.167885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.167900 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.167910 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.168705 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.168741 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.168750 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.168766 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.168775 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: E0930 14:20:35.180911 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:35Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.184726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.184806 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.184826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.184843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.184854 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: E0930 14:20:35.195931 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:35Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.198882 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.198910 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.198920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.198934 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.198942 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: E0930 14:20:35.213500 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:35Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.218284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.218313 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.218322 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.218334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.218343 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: E0930 14:20:35.230122 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:35Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.236089 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.236118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.236127 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.236143 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.236155 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: E0930 14:20:35.247326 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:35Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:35 crc kubenswrapper[4799]: E0930 14:20:35.247472 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.269953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.269986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.269996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.270011 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.270023 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.372450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.372515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.372529 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.372546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.372555 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.475306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.475336 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.475346 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.475362 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.475372 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.503039 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.503086 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:35 crc kubenswrapper[4799]: E0930 14:20:35.503191 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:35 crc kubenswrapper[4799]: E0930 14:20:35.503279 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.578264 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.578315 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.578326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.578340 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.578366 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.681047 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.681077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.681085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.681099 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.681107 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.784560 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.784602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.784613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.784630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.784642 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.886499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.886544 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.886556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.886573 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.886583 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.989389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.989434 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.989444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.989461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:35 crc kubenswrapper[4799]: I0930 14:20:35.989472 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:35Z","lastTransitionTime":"2025-09-30T14:20:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.091442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.091479 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.091488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.091503 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.091511 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.193458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.193500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.193513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.193526 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.193536 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.296211 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.296243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.296252 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.296264 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.296272 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.398795 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.398838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.398850 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.398870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.398881 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.501271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.501323 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.501334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.501350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.501366 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.502505 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:36 crc kubenswrapper[4799]: E0930 14:20:36.502621 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.502696 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:36 crc kubenswrapper[4799]: E0930 14:20:36.502805 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.503365 4799 scope.go:117] "RemoveContainer" containerID="af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32" Sep 30 14:20:36 crc kubenswrapper[4799]: E0930 14:20:36.503529 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.603802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.603836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.603847 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.603863 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.603874 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.705786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.705815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.705823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.705836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.705845 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.807954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.807981 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.807990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.808005 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.808013 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.910463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.910508 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.910524 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.910542 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:36 crc kubenswrapper[4799]: I0930 14:20:36.910553 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:36Z","lastTransitionTime":"2025-09-30T14:20:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.012499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.012526 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.012534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.012551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.012575 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.114437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.114471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.114480 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.114493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.114530 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.217708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.217757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.217783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.217804 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.217819 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.320045 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.320132 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.320144 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.320159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.320167 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.421928 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.421972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.421983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.422000 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.422012 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.502469 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.502542 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:37 crc kubenswrapper[4799]: E0930 14:20:37.502604 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:37 crc kubenswrapper[4799]: E0930 14:20:37.502820 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.524499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.524538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.524551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.524568 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.524582 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.626910 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.626940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.626948 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.626963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.626972 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.729162 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.729196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.729205 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.729223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.729232 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.830974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.831022 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.831039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.831054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.831064 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.933123 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.933157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.933165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.933183 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:37 crc kubenswrapper[4799]: I0930 14:20:37.933200 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:37Z","lastTransitionTime":"2025-09-30T14:20:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.036168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.036210 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.036226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.036247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.036261 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.139762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.139799 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.139807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.139824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.139833 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.242354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.242391 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.242401 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.242417 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.242427 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.344525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.344557 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.344564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.344577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.344587 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.446070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.446101 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.446110 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.446123 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.446133 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.507858 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:38 crc kubenswrapper[4799]: E0930 14:20:38.508017 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.508792 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:38 crc kubenswrapper[4799]: E0930 14:20:38.508923 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.531621 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.549214 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.549417 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.549458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.549473 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.549494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.549508 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.565040 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.583545 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.602287 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.625339 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.639248 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.653915 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.653953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.653961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.653976 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.653987 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.655388 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.668721 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.682715 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.693815 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.709498 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.725167 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.738178 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.751308 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.757510 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.757575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.757587 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.757628 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.757643 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.768663 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.782895 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.794755 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:38Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.859704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.859737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.859745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.859759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.859768 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.962312 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.962639 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.962786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.962921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:38 crc kubenswrapper[4799]: I0930 14:20:38.963057 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:38Z","lastTransitionTime":"2025-09-30T14:20:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.065329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.065374 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.065391 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.065411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.065424 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.168195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.168449 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.168522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.168602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.168703 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.275119 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.275158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.275166 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.275212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.275231 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.377730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.377949 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.378007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.378070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.378133 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.480360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.480598 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.480704 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.480781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.480842 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.502821 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:39 crc kubenswrapper[4799]: E0930 14:20:39.502967 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.503035 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:39 crc kubenswrapper[4799]: E0930 14:20:39.503196 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.582804 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.582835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.582846 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.582871 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.582881 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.685604 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.685639 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.685668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.685686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.685695 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.788453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.788561 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.788575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.788591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.788601 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.890270 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.890305 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.890314 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.890338 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.890346 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.992555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.992583 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.992592 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.992605 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:39 crc kubenswrapper[4799]: I0930 14:20:39.992613 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:39Z","lastTransitionTime":"2025-09-30T14:20:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.094920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.094967 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.094979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.094996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.095016 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.197133 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.197168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.197176 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.197190 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.197199 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.299882 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.299910 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.299927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.299950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.299964 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.402979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.403033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.403043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.403066 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.403079 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.504713 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:40 crc kubenswrapper[4799]: E0930 14:20:40.504827 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.505019 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:40 crc kubenswrapper[4799]: E0930 14:20:40.505087 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.507877 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.507906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.507914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.507927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.507937 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.610539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.610578 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.610586 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.610603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.610615 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.715538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.716033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.716045 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.716070 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.716082 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.818766 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.818805 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.818818 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.818844 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.818858 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.921267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.921297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.921306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.921318 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:40 crc kubenswrapper[4799]: I0930 14:20:40.921327 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:40Z","lastTransitionTime":"2025-09-30T14:20:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.023404 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.023448 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.023458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.023473 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.023483 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.125949 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.125995 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.126008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.126029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.126052 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.228399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.228425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.228433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.228446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.228456 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.331249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.331276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.331284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.331297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.331306 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.433875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.433905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.433914 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.433927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.433937 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.503134 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:41 crc kubenswrapper[4799]: E0930 14:20:41.503255 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.503455 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:41 crc kubenswrapper[4799]: E0930 14:20:41.503514 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.536778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.536835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.536848 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.536864 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.537257 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.639639 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.639703 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.639713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.639729 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.639739 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.741962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.741992 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.742002 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.742017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.742029 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.844290 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.844591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.844675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.844778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.844840 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.947109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.947160 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.947173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.947195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:41 crc kubenswrapper[4799]: I0930 14:20:41.947208 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:41Z","lastTransitionTime":"2025-09-30T14:20:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.049382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.049415 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.049427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.049443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.049455 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.152027 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.152086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.152099 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.152116 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.152127 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.254918 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.255204 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.255273 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.255339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.255425 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.358109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.358143 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.358153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.358168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.358180 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.460656 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.460904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.461001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.461067 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.461135 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.502426 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.502558 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:42 crc kubenswrapper[4799]: E0930 14:20:42.502775 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:42 crc kubenswrapper[4799]: E0930 14:20:42.502984 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.563514 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.563582 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.563594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.563616 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.563627 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.666382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.666440 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.666453 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.666472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.666484 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.769684 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.769724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.769735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.769755 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.769766 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.872101 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.872139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.872149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.872164 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.872175 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.974489 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.974539 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.974553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.974571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:42 crc kubenswrapper[4799]: I0930 14:20:42.974583 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:42Z","lastTransitionTime":"2025-09-30T14:20:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.076773 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.076823 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.076836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.076853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.076867 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.179956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.179993 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.180003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.180019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.180029 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.282285 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.282319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.282327 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.282344 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.282355 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.384599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.384858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.384940 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.385021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.385136 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.487589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.487638 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.487648 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.487679 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.487691 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.502527 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:43 crc kubenswrapper[4799]: E0930 14:20:43.502638 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.502541 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:43 crc kubenswrapper[4799]: E0930 14:20:43.503067 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.589946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.589979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.589990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.590003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.590012 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.692597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.693091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.693181 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.693252 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.693316 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.795989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.796244 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.796324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.796410 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.796495 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.899004 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.899051 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.899062 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.899078 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:43 crc kubenswrapper[4799]: I0930 14:20:43.899089 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:43Z","lastTransitionTime":"2025-09-30T14:20:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.001308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.001517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.001612 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.001711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.001779 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.104208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.104254 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.104265 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.104279 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.104287 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.206746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.206782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.206792 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.206807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.206818 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.309024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.309055 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.309063 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.309077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.309086 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.410947 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.410983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.410994 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.411007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.411017 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.519530 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:44 crc kubenswrapper[4799]: E0930 14:20:44.520660 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.520136 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.521849 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.521929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.522002 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.522063 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: E0930 14:20:44.521023 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:44 crc kubenswrapper[4799]: E0930 14:20:44.522219 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs podName:15ddd33f-0b7f-4cb5-b09a-03c211c759c8 nodeName:}" failed. No retries permitted until 2025-09-30 14:21:16.522199687 +0000 UTC m=+98.605800114 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs") pod "network-metrics-daemon-hghbr" (UID: "15ddd33f-0b7f-4cb5-b09a-03c211c759c8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.520949 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.519634 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:44 crc kubenswrapper[4799]: E0930 14:20:44.522330 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.624822 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.625168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.625241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.625311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.625365 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.727719 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.727770 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.727783 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.727803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.727815 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.830223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.830272 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.830313 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.830334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.830348 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.933447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.933730 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.933802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.933872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:44 crc kubenswrapper[4799]: I0930 14:20:44.933942 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:44Z","lastTransitionTime":"2025-09-30T14:20:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.036154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.036411 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.036475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.036547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.036620 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.139292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.139324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.139335 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.139352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.139362 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.242076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.242118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.242130 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.242146 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.242158 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.344126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.344164 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.344173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.344186 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.344196 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.446802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.446843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.446852 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.446866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.446875 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.502976 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.502976 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:45 crc kubenswrapper[4799]: E0930 14:20:45.503147 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:45 crc kubenswrapper[4799]: E0930 14:20:45.503075 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.549499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.549538 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.549549 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.549568 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.549581 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.646467 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.646545 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.646555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.646570 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.646578 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: E0930 14:20:45.659537 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:45Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.662928 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.662960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.662969 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.662986 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.662999 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: E0930 14:20:45.677442 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:45Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.681376 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.681426 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.681469 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.681488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.681498 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: E0930 14:20:45.695438 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:45Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.699780 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.699830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.699843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.699864 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.699876 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: E0930 14:20:45.719954 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:45Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.723445 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.723476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.723488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.723507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.723518 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: E0930 14:20:45.735787 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:45Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:45 crc kubenswrapper[4799]: E0930 14:20:45.735902 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.737457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.737488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.737497 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.737513 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.737524 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.840266 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.840327 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.840340 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.840361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.840371 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.944085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.944179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.944208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.944246 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:45 crc kubenswrapper[4799]: I0930 14:20:45.944273 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:45Z","lastTransitionTime":"2025-09-30T14:20:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.046964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.047236 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.047297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.047368 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.047445 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.149726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.149759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.149768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.149781 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.149790 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.252369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.252432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.252445 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.252468 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.252482 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.354690 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.354737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.354749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.354771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.354784 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.457907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.457962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.457976 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.458002 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.458018 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.502290 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.502372 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:46 crc kubenswrapper[4799]: E0930 14:20:46.502504 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:46 crc kubenswrapper[4799]: E0930 14:20:46.502712 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.560890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.560935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.560950 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.560971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.560987 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.664102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.664157 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.664172 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.664195 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.664208 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.767798 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.767851 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.767865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.767912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.767925 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.870537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.870854 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.870962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.871048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.871115 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.974187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.974235 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.974249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.974265 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:46 crc kubenswrapper[4799]: I0930 14:20:46.974277 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:46Z","lastTransitionTime":"2025-09-30T14:20:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.077092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.077135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.077148 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.077163 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.077176 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.179378 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.179840 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.179925 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.180001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.180076 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.282987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.283025 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.283035 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.283056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.283068 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.385896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.386301 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.386437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.386561 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.386627 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.489582 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.489640 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.489683 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.489702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.489721 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.503059 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.503154 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:47 crc kubenswrapper[4799]: E0930 14:20:47.503482 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:47 crc kubenswrapper[4799]: E0930 14:20:47.503411 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.592849 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.592906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.592917 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.592942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.592957 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.695827 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.695874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.695888 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.695906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.695919 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.798528 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.799292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.799384 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.799492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.799584 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.902241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.902739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.902820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.902904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.902967 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:47Z","lastTransitionTime":"2025-09-30T14:20:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.908868 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/0.log" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.909011 4799 generic.go:334] "Generic (PLEG): container finished" podID="5517db4f-9237-4b92-a672-6fa7afcd7e3f" containerID="f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de" exitCode=1 Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.909103 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sjjm8" event={"ID":"5517db4f-9237-4b92-a672-6fa7afcd7e3f","Type":"ContainerDied","Data":"f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de"} Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.909606 4799 scope.go:117] "RemoveContainer" containerID="f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.926919 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:47Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.943099 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:47Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.959620 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"2025-09-30T14:20:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754\\\\n2025-09-30T14:20:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754 to /host/opt/cni/bin/\\\\n2025-09-30T14:20:02Z [verbose] multus-daemon started\\\\n2025-09-30T14:20:02Z [verbose] Readiness Indicator file check\\\\n2025-09-30T14:20:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:47Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.979405 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:47Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:47 crc kubenswrapper[4799]: I0930 14:20:47.995737 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:47Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.004995 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.005047 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.005058 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.005076 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.005108 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.010453 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.027134 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.046132 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.075752 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.102389 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.106796 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.106829 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.106838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.106853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.106861 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.144361 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.161080 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.182181 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.193113 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.208948 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.209003 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.209016 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.209037 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.209049 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.214362 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.228014 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.242951 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.257181 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.311033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.311072 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.311081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.311121 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.311131 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.413867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.413919 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.413930 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.413951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.413976 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.502330 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.502369 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:48 crc kubenswrapper[4799]: E0930 14:20:48.502514 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:48 crc kubenswrapper[4799]: E0930 14:20:48.502606 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.515779 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.515825 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.515838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.515855 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.515867 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.518541 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.533001 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.545276 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.568022 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.586033 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.607808 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.618552 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.618617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.618631 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.618689 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.618706 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.623221 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.641534 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.660847 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.676486 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.694011 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.709076 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.723297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.723342 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.723353 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.723373 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.723387 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.725137 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"2025-09-30T14:20:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754\\\\n2025-09-30T14:20:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754 to /host/opt/cni/bin/\\\\n2025-09-30T14:20:02Z [verbose] multus-daemon started\\\\n2025-09-30T14:20:02Z [verbose] Readiness Indicator file check\\\\n2025-09-30T14:20:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.737523 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.751534 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.765234 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.778896 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.794278 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.826056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.826118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.826132 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.826153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.826166 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.914186 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/0.log" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.914244 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sjjm8" event={"ID":"5517db4f-9237-4b92-a672-6fa7afcd7e3f","Type":"ContainerStarted","Data":"f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.928346 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.928624 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.928650 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.928675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.928692 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.928702 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:48Z","lastTransitionTime":"2025-09-30T14:20:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.937741 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.950119 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.971700 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:48 crc kubenswrapper[4799]: I0930 14:20:48.986918 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.001097 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:48Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.013885 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.029649 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.030463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.030488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.030495 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.030511 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.030524 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.049295 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.060018 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.071605 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.085373 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.097062 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"2025-09-30T14:20:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754\\\\n2025-09-30T14:20:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754 to /host/opt/cni/bin/\\\\n2025-09-30T14:20:02Z [verbose] multus-daemon started\\\\n2025-09-30T14:20:02Z [verbose] Readiness Indicator file check\\\\n2025-09-30T14:20:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.106544 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.116687 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.128844 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.132935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.132972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.132982 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.132999 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.133023 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.141630 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.153671 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:49Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.235980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.236029 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.236041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.236058 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.236070 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.338300 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.338337 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.338347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.338361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.338370 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.440402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.440452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.440460 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.440474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.440483 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.503088 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.503112 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:49 crc kubenswrapper[4799]: E0930 14:20:49.503246 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:49 crc kubenswrapper[4799]: E0930 14:20:49.503352 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.543001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.543032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.543041 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.543058 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.543070 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.646198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.646245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.646257 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.646276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.646289 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.749906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.749955 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.749966 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.749988 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.750002 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.852996 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.853042 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.853056 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.853085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.853099 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.955883 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.955913 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.955922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.955937 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:49 crc kubenswrapper[4799]: I0930 14:20:49.955946 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:49Z","lastTransitionTime":"2025-09-30T14:20:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.058102 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.058133 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.058141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.058155 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.058164 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.167072 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.167150 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.167164 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.167187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.167203 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.269809 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.269847 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.269858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.269873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.269885 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.372082 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.372114 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.372126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.372141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.372151 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.474341 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.474369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.474379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.474393 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.474402 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.502130 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.502226 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:50 crc kubenswrapper[4799]: E0930 14:20:50.502372 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:50 crc kubenswrapper[4799]: E0930 14:20:50.502536 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.577689 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.577749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.577760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.577775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.577809 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.680320 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.680359 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.680370 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.680388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.680400 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.782183 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.782229 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.782241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.782258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.782270 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.884948 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.885007 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.885019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.885036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.885048 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.986780 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.986812 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.986821 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.986835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:50 crc kubenswrapper[4799]: I0930 14:20:50.986843 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:50Z","lastTransitionTime":"2025-09-30T14:20:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.089600 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.089675 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.089686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.089702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.089712 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.193755 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.193830 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.193846 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.193868 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.193880 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.296575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.296618 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.296630 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.296672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.296682 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.399458 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.399516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.399533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.399550 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.399561 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.501472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.501506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.501515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.501528 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.501613 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.502915 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.503109 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:51 crc kubenswrapper[4799]: E0930 14:20:51.503170 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:51 crc kubenswrapper[4799]: E0930 14:20:51.503568 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.503689 4799 scope.go:117] "RemoveContainer" containerID="af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.605369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.605437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.605459 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.605493 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.605512 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.708735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.708775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.708787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.708809 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.708823 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.812500 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.812774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.812842 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.812920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.812983 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.915945 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.915987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.915997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.916013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.916025 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:51Z","lastTransitionTime":"2025-09-30T14:20:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.925177 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/2.log" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.928359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.928859 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.945083 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:51Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.972570 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:51Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:51 crc kubenswrapper[4799]: I0930 14:20:51.988851 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"2025-09-30T14:20:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754\\\\n2025-09-30T14:20:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754 to /host/opt/cni/bin/\\\\n2025-09-30T14:20:02Z [verbose] multus-daemon started\\\\n2025-09-30T14:20:02Z [verbose] Readiness Indicator file check\\\\n2025-09-30T14:20:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:51Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.000978 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:51Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.018395 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.018931 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.018974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.018988 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.019015 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.019033 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.033161 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.049005 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.065677 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.081818 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.096155 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.109471 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.121938 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.122162 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.122282 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.122377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.122455 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.129968 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.145043 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.169891 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.186543 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.204933 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.218672 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.227165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.227220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.227230 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.227249 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.227262 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.235672 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.329868 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.329917 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.329933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.329953 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.329964 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.432742 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.432772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.432780 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.432793 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.432801 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.504696 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:52 crc kubenswrapper[4799]: E0930 14:20:52.504815 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.504975 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:52 crc kubenswrapper[4799]: E0930 14:20:52.505025 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.534702 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.534732 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.534751 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.534763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.534772 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.637636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.637689 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.637699 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.637715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.637726 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.740937 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.740975 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.740984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.740998 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.741006 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.843427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.843477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.843492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.843507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.843517 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.932339 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/3.log" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.932900 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/2.log" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.935955 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" exitCode=1 Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.936004 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.936041 4799 scope.go:117] "RemoveContainer" containerID="af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.936590 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:20:52 crc kubenswrapper[4799]: E0930 14:20:52.936829 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.946334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.946375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.946385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.946456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.946473 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:52Z","lastTransitionTime":"2025-09-30T14:20:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.953215 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.966218 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.979788 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:52 crc kubenswrapper[4799]: I0930 14:20:52.991684 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:52Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.002808 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.015861 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.027118 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.046975 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.048439 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.048484 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.048496 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.048515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.048527 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.059399 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.077375 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.094171 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af6d52df71672a237b530e82e81c9c66d83e2917f2c2cc14e1e9fce3bcc71e32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:24Z\\\",\\\"message\\\":\\\" handler 5\\\\nI0930 14:20:24.438477 6368 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0930 14:20:24.438483 6368 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0930 14:20:24.438525 6368 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.438941 6368 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439043 6368 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439160 6368 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0930 14:20:24.439344 6368 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439548 6368 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0930 14:20:24.439709 6368 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:52Z\\\",\\\"message\\\":\\\"ft-image-registry/node-ca-vmvlc openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/machine-config-daemon-wr2vb openshift-multus/network-metrics-daemon-hghbr openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-etcd/etcd-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-hstq9]\\\\nI0930 14:20:52.378887 6714 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0930 14:20:52.378902 6714 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nI0930 14:20:52.378913 6714 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nI0930 14:20:52.378924 6714 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hstq9 in node crc\\\\nI0930 14:20:52.378929 6714 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9 after 0 failed attempt(s)\\\\nI0930 14:20:52.378936 6714 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nF0930 14:20:52.378952 6714 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.107491 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.127566 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.139868 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.151018 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.151073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.151084 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.151104 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.151123 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.151679 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"2025-09-30T14:20:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754\\\\n2025-09-30T14:20:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754 to /host/opt/cni/bin/\\\\n2025-09-30T14:20:02Z [verbose] multus-daemon started\\\\n2025-09-30T14:20:02Z [verbose] Readiness Indicator file check\\\\n2025-09-30T14:20:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.163013 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.174245 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.185369 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.253672 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.253714 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.253725 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.253746 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.253759 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.356935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.356967 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.356976 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.356989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.356998 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.459237 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.459276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.459284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.459298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.459308 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.502886 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.503001 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:53 crc kubenswrapper[4799]: E0930 14:20:53.503297 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:53 crc kubenswrapper[4799]: E0930 14:20:53.503366 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.513067 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.562344 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.562389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.562400 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.562417 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.562430 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.665138 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.665173 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.665180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.665194 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.665203 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.768705 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.768768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.768788 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.768817 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.768836 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.871764 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.871839 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.871863 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.871899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.871921 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.943248 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/3.log" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.947588 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:20:53 crc kubenswrapper[4799]: E0930 14:20:53.947785 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.972104 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.975361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.975429 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.975450 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.975482 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.975510 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:53Z","lastTransitionTime":"2025-09-30T14:20:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:53 crc kubenswrapper[4799]: I0930 14:20:53.991373 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:53Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.006913 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.021605 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.042632 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.063352 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:52Z\\\",\\\"message\\\":\\\"ft-image-registry/node-ca-vmvlc openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/machine-config-daemon-wr2vb openshift-multus/network-metrics-daemon-hghbr openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-etcd/etcd-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-hstq9]\\\\nI0930 14:20:52.378887 6714 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0930 14:20:52.378902 6714 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nI0930 14:20:52.378913 6714 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nI0930 14:20:52.378924 6714 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hstq9 in node crc\\\\nI0930 14:20:52.378929 6714 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9 after 0 failed attempt(s)\\\\nI0930 14:20:52.378936 6714 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nF0930 14:20:52.378952 6714 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.077912 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.079369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.079418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.079430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.079451 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.079466 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.093119 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.109590 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.130396 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"2025-09-30T14:20:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754\\\\n2025-09-30T14:20:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754 to /host/opt/cni/bin/\\\\n2025-09-30T14:20:02Z [verbose] multus-daemon started\\\\n2025-09-30T14:20:02Z [verbose] Readiness Indicator file check\\\\n2025-09-30T14:20:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.142706 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.161565 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.179795 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.183033 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.183081 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.183092 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.183108 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.183119 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.196785 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.212095 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.228791 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.243607 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.257883 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4833d883-8f06-4e1f-bed3-3dc153f504cc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31b21f4329ffe89a3a4fa23f7fd3d5168eebc20e1849632375c40710b11abc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05566b59254df7f2aa21029d8333078db94541ab7e274195010613cad824e85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05566b59254df7f2aa21029d8333078db94541ab7e274195010613cad824e85e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.272051 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:54Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.286506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.286595 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.286612 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.286636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.286715 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.390802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.390871 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.390898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.390929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.390951 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.493827 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.493875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.493885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.493905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.493918 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.502755 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.502865 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:54 crc kubenswrapper[4799]: E0930 14:20:54.503551 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:54 crc kubenswrapper[4799]: E0930 14:20:54.503789 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.596853 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.596896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.596907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.596927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.596939 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.700306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.700865 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.701038 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.701182 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.701300 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.804162 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.804209 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.804218 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.804234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.804246 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.907222 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.907271 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.907282 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.907299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:54 crc kubenswrapper[4799]: I0930 14:20:54.907310 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:54Z","lastTransitionTime":"2025-09-30T14:20:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.011159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.011231 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.011245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.011276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.011290 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.114030 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.114097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.114113 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.114137 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.114154 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.216611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.216669 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.216679 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.216695 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.216704 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.319486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.319534 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.319547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.319566 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.319581 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.423085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.423140 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.423153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.423174 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.423185 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.502853 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.502942 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:55 crc kubenswrapper[4799]: E0930 14:20:55.503030 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:55 crc kubenswrapper[4799]: E0930 14:20:55.503239 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.525872 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.525905 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.525912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.525926 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.525935 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.628637 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.628715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.628728 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.628751 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.628766 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.731720 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.731766 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.731776 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.731792 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.731802 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.834535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.834569 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.834578 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.834624 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.834635 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.938088 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.938149 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.938168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.938198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:55 crc kubenswrapper[4799]: I0930 14:20:55.938217 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:55Z","lastTransitionTime":"2025-09-30T14:20:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.008103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.008140 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.008151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.008168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.008178 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: E0930 14:20:56.022531 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:56Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.027146 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.027198 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.027226 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.027241 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.027253 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: E0930 14:20:56.043551 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:56Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.048221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.048284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.048298 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.048317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.048329 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: E0930 14:20:56.062635 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:56Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.067343 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.067372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.067380 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.067399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.067411 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: E0930 14:20:56.081368 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:56Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.084881 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.084903 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.084912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.084927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.084936 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: E0930 14:20:56.098508 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:56Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:56 crc kubenswrapper[4799]: E0930 14:20:56.098621 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.100168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.100194 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.100205 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.100220 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.100229 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.202928 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.202968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.202980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.202997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.203008 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.306080 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.306146 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.306159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.306178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.306192 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.409106 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.409161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.409172 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.409194 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.409208 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.503191 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.503260 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:56 crc kubenswrapper[4799]: E0930 14:20:56.503348 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:56 crc kubenswrapper[4799]: E0930 14:20:56.503526 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.511509 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.511546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.511556 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.511571 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.511583 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.613972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.614012 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.614021 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.614035 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.614045 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.716094 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.716126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.716138 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.716153 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.716161 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.818610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.818665 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.818674 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.818690 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.818700 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.921276 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.921319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.921345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.921363 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:56 crc kubenswrapper[4799]: I0930 14:20:56.921372 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:56Z","lastTransitionTime":"2025-09-30T14:20:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.023477 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.023523 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.023533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.023547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.023557 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.126352 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.126421 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.126432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.126447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.126455 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.228916 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.228990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.229009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.229039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.229057 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.331629 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.331718 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.331733 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.331751 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.331762 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.434390 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.434441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.434454 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.434475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.434485 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.502935 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.503000 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:57 crc kubenswrapper[4799]: E0930 14:20:57.503071 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:57 crc kubenswrapper[4799]: E0930 14:20:57.503157 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.537546 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.537612 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.537623 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.537642 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.537686 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.640425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.640463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.640472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.640486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.640497 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.743032 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.743075 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.743086 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.743103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.743115 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.847966 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.848028 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.848043 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.848064 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.848080 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.951267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.951318 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.951330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.951349 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:57 crc kubenswrapper[4799]: I0930 14:20:57.951360 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:57Z","lastTransitionTime":"2025-09-30T14:20:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.054471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.054524 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.054535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.054553 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.054567 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.157682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.157747 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.157758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.157778 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.157788 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.260676 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.260739 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.260749 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.260768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.260778 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.364052 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.364112 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.364127 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.364151 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.364171 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.467166 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.467229 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.467243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.467268 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.467286 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.503106 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.503267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:20:58 crc kubenswrapper[4799]: E0930 14:20:58.503408 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:20:58 crc kubenswrapper[4799]: E0930 14:20:58.503552 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.517489 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.528668 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cad3b181-b2f5-4961-8134-f471bff50c40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aeb596116175d6f778d0f27e8a5e93d02530cb9772edeb3779ea197809970f9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g747d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-wr2vb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.545630 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hstq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cce844e9-25b0-49e7-9ca3-899b986d0f6f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89217018942bd47d33ab65483a281c0a475205400db8ea7de626ef7d36fbcd36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8d68fb5ae251931d490b11a2831ec5c765147742726d88e24ce9c6ed761429f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94a02c350fa6f9f581fdcf6e2b9c539180ce1610523d6322ba9291b9b17f27e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ecc8f0afb98fa8991cf42807917943e3a703240240ea5791f9f92e05de4457b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61fb7564f5e07cbdf88b73e96d29c8c34a7c4a9cb73dcfc8d53446e54b07603f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fdbcd4ac9bdf9a82ca483463f62b5aade5711018fc9ed6cb63187ce48c081927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2586f62465270e776757e9e44467ba0b69c30aa01447e8838461267ae3188947\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hgb7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hstq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.564338 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:52Z\\\",\\\"message\\\":\\\"ft-image-registry/node-ca-vmvlc openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/machine-config-daemon-wr2vb openshift-multus/network-metrics-daemon-hghbr openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-etcd/etcd-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-hstq9]\\\\nI0930 14:20:52.378887 6714 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0930 14:20:52.378902 6714 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nI0930 14:20:52.378913 6714 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nI0930 14:20:52.378924 6714 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hstq9 in node crc\\\\nI0930 14:20:52.378929 6714 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9 after 0 failed attempt(s)\\\\nI0930 14:20:52.378936 6714 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nF0930 14:20:52.378952 6714 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.569533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.569609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.569627 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.569673 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.569696 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.577233 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.600197 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.611591 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.624245 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-sjjm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5517db4f-9237-4b92-a672-6fa7afcd7e3f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:47Z\\\",\\\"message\\\":\\\"2025-09-30T14:20:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754\\\\n2025-09-30T14:20:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_745f4ef1-4e4f-4384-a319-00785f92f754 to /host/opt/cni/bin/\\\\n2025-09-30T14:20:02Z [verbose] multus-daemon started\\\\n2025-09-30T14:20:02Z [verbose] Readiness Indicator file check\\\\n2025-09-30T14:20:47Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5tsx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-multus\"/\"multus-sjjm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.635466 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hlg57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b1e4d60-4242-4ca8-a5c4-d76dd6f0046d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c124b2a8e61c1abddfcaeeb1be798ce47aa3f99594c20527076824884f19787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lwl67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hlg57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.647087 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.658032 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.669714 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d16ad3a5-b9c0-42a6-aff2-11b58fb566fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b16e4beedb5e1c971675c5ca837a71770331b6a5e4d4eda99bf165f5e15a1e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://082030ac53cf3685194e2a65199d9a0494072bd7a66e09f55f24b52b5968a48f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wgr2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7vhf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.671861 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.671890 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.671901 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.671919 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.671930 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.682704 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hghbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bkzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:12Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hghbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.695597 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b428a6b92a1c591252f9b635879bb4bae0f2d15237a571b08aca1c74687ddc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.706843 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.716145 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4833d883-8f06-4e1f-bed3-3dc153f504cc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31b21f4329ffe89a3a4fa23f7fd3d5168eebc20e1849632375c40710b11abc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05566b59254df7f2aa21029d8333078db94541ab7e274195010613cad824e85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05566b59254df7f2aa21029d8333078db94541ab7e274195010613cad824e85e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.726700 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dcdb9686a4175681fb8a3dddbdfd0cb7f51303935aa0343a8bbffcf0561a51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.740408 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24cd5d87-672a-4a8f-8a01-29703c892bc0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b820d82b26559f34f0a844d5f10d5a78e8903633f51a00c58faf8542976ee601\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://276799abdfed93f0185e2b844868fdc692c2c374144f59b83312b841893f8b37\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52445bafca0732f4b4570c54fe97a162160807b9bbce69c0e072f123fe2d755\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4341957cf8449f848043024bee3ddcae62a3f5c8945a1d8e731fdb7feef1256\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c042510e3a08dfce962f07bcfb6376b60136fbb02e58ab64a88976d84093ab6c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-30T14:19:57Z\\\",\\\"message\\\":\\\"mespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.596459 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0930 14:19:57.597024 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3422204409/tls.crt::/tmp/serving-cert-3422204409/tls.key\\\\\\\"\\\\nI0930 14:19:57.743245 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0930 14:19:57.747157 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0930 14:19:57.747178 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0930 14:19:57.747194 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0930 14:19:57.747199 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0930 14:19:57.754204 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0930 14:19:57.754239 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0930 14:19:57.754247 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0930 14:19:57.754257 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0930 14:19:57.754260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0930 14:19:57.754262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0930 14:19:57.754265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0930 14:19:57.756750 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0930 14:19:57.756791 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61afecf82dfa83747e33a0ea6b02d18d0ff466d3dc69804c5e807ac44503bad1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fedc7b0aa93ced92e2d89db6b09cad3db9368fbb9bcae9b7438dae08f221d4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.752325 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a7239e50-f4e3-45af-a6a6-63396d22088b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9df76d18e5aa14056dc22e5096f0565e00124cbc736f0e34a4b9065ffaae4265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8da3b9c3db35fb92679b1db3a03b5fbc9303db22dff11ca516423d81ae46fca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0785580d01a4fff4c682b013b66da6ed198702b9f603607cd8e7db86f91803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bd691293318baba8e4f3e0ce1f9a52bb0e84f87aebdc4934a4e29cf64e9d063\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:20:58Z is after 2025-08-24T17:21:41Z" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.774284 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.774317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.774326 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.774339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.774348 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.877245 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.877283 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.877292 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.877309 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.877319 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.979921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.980223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.980293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.980367 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:58 crc kubenswrapper[4799]: I0930 14:20:58.980431 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:58Z","lastTransitionTime":"2025-09-30T14:20:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.083274 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.083318 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.083329 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.083345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.083356 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.186048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.186104 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.186130 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.186156 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.186171 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.288491 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.288525 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.288537 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.288551 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.288563 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.390864 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.390909 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.390925 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.390943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.390955 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.492506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.492540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.492548 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.492561 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.492571 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.502727 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.502779 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:20:59 crc kubenswrapper[4799]: E0930 14:20:59.502873 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:20:59 crc kubenswrapper[4799]: E0930 14:20:59.502998 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.596201 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.596244 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.596262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.596281 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.596295 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.698773 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.698813 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.698825 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.698841 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.698853 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.800670 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.800713 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.800722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.800920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.800931 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.902775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.902836 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.902846 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.902858 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:20:59 crc kubenswrapper[4799]: I0930 14:20:59.902868 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:20:59Z","lastTransitionTime":"2025-09-30T14:20:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.005128 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.005165 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.005177 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.005193 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.005209 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.107690 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.107863 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.107878 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.107896 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.107908 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.211178 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.211221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.211233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.211251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.211265 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.313264 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.313299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.313308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.313321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.313329 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.416065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.416100 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.416120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.416135 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.416145 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.502892 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:00 crc kubenswrapper[4799]: E0930 14:21:00.503080 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.503234 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:00 crc kubenswrapper[4799]: E0930 14:21:00.503502 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.518293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.518740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.518826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.518921 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.518992 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.621954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.622011 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.622049 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.622077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.622087 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.725230 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.725565 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.725715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.725874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.725963 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.829409 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.829475 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.829492 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.829517 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.829534 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.933315 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.933859 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.933991 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.934080 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:00 crc kubenswrapper[4799]: I0930 14:21:00.934185 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:00Z","lastTransitionTime":"2025-09-30T14:21:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.036572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.036820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.036897 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.037006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.037083 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.140457 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.140503 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.140516 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.140533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.140544 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.242958 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.242997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.243006 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.243018 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.243027 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.346248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.346293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.346308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.346325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.346337 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.448682 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.449024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.449034 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.449048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.449061 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.502075 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.502179 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:01 crc kubenswrapper[4799]: E0930 14:21:01.502425 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:01 crc kubenswrapper[4799]: E0930 14:21:01.502217 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.551382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.551430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.551441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.551536 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.551551 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.653718 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.653758 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.653768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.653782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.653791 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.756902 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.756964 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.756974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.756989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.756997 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.859262 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.859299 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.859308 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.859321 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.859329 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.961606 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.961644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.961667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.961683 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:01 crc kubenswrapper[4799]: I0930 14:21:01.961696 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:01Z","lastTransitionTime":"2025-09-30T14:21:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.064700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.064737 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.064745 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.064760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.064769 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.167048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.167085 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.167093 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.167116 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.167134 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.269977 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.270020 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.270031 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.270048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.270059 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.373328 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.373401 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.373417 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.373436 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.373474 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.476543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.476639 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.476721 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.476762 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.476786 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.502221 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.502340 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:02 crc kubenswrapper[4799]: E0930 14:21:02.502395 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:02 crc kubenswrapper[4799]: E0930 14:21:02.502588 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.579310 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.579369 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.579381 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.579402 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.579420 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.682577 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.682627 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.682636 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.682681 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.682692 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.785333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.785385 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.785397 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.785415 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.785428 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.887405 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.887437 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.887446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.887459 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.887467 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.989670 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.989712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.989722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.989738 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:02 crc kubenswrapper[4799]: I0930 14:21:02.989748 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:02Z","lastTransitionTime":"2025-09-30T14:21:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.091687 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.091723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.091731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.091744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.091753 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.194159 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.194200 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.194208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.194223 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.194232 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.295946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.295988 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.295998 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.296014 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.296024 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.398302 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.398338 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.398350 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.398365 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.398376 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.428737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.428877 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.429034 4799 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.429100 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.429079703 +0000 UTC m=+149.512680130 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.429138 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.429128084 +0000 UTC m=+149.512728511 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.500708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.500768 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.500787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.500810 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.500826 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.502791 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.502807 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.502883 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.502995 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.529427 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.529472 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.529494 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529678 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529709 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529721 4799 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529755 4799 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529782 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.529764503 +0000 UTC m=+149.613364930 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529814 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.529798693 +0000 UTC m=+149.613399120 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529832 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529882 4799 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529903 4799 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:21:03 crc kubenswrapper[4799]: E0930 14:21:03.529994 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.529967547 +0000 UTC m=+149.613568004 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.603208 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.603247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.603258 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.603275 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.603286 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.705531 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.705564 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.705572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.705585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.705601 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.808307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.808348 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.808360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.808377 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.808388 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.910141 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.910186 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.910197 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.910214 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:03 crc kubenswrapper[4799]: I0930 14:21:03.910228 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:03Z","lastTransitionTime":"2025-09-30T14:21:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.012330 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.012422 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.012442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.012466 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.012482 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.115126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.115169 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.115180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.115196 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.115206 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.217468 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.217505 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.217515 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.217533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.217546 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.320427 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.320804 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.320951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.321091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.321210 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.423558 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.423599 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.423611 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.423626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.423635 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.503159 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.503159 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:04 crc kubenswrapper[4799]: E0930 14:21:04.503282 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:04 crc kubenswrapper[4799]: E0930 14:21:04.503358 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.525603 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.525671 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.525691 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.525710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.525720 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.627266 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.627300 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.627313 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.627334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.627344 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.729933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.729966 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.729976 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.729990 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.729998 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.832960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.833036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.833050 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.833065 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.833076 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.936036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.936073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.936083 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.936097 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:04 crc kubenswrapper[4799]: I0930 14:21:04.936106 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:04Z","lastTransitionTime":"2025-09-30T14:21:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.038291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.038324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.038335 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.038362 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.038374 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.140576 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.140617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.140629 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.140668 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.140681 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.243572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.243618 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.243629 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.243644 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.243676 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.346060 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.346116 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.346134 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.346158 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.346175 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.448772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.448799 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.448807 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.448837 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.448846 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.502888 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:05 crc kubenswrapper[4799]: E0930 14:21:05.503010 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.503210 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:05 crc kubenswrapper[4799]: E0930 14:21:05.503276 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.503976 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:21:05 crc kubenswrapper[4799]: E0930 14:21:05.504173 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.551740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.551774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.551784 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.551799 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.551810 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.653883 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.653929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.653942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.653959 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.654197 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.756216 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.756275 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.756286 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.756304 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.756316 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.858594 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.858626 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.858634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.858670 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.858680 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.961591 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.961883 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.961954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.962013 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:05 crc kubenswrapper[4799]: I0930 14:21:05.962076 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:05Z","lastTransitionTime":"2025-09-30T14:21:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.064394 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.064432 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.064442 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.064455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.064464 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.166690 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.166735 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.166744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.166757 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.166767 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.184246 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.184289 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.184297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.184311 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.184322 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: E0930 14:21:06.196198 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.200431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.200463 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.200498 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.200518 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.200531 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: E0930 14:21:06.212438 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.216291 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.216333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.216347 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.216362 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.216372 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: E0930 14:21:06.228021 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.231399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.231430 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.231441 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.231456 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.231466 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: E0930 14:21:06.243729 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.251760 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.251793 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.251802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.251815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.251824 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: E0930 14:21:06.263371 4799 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-30T14:21:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"78226594-f56c-4f63-88aa-8c08111755bf\\\",\\\"systemUUID\\\":\\\"86f82190-ecf6-4439-9be4-52d5a249c8f1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:06Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:06 crc kubenswrapper[4799]: E0930 14:21:06.263483 4799 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.268956 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.269170 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.269179 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.269191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.269200 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.371264 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.371297 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.371307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.371324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.371335 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.473339 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.473408 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.473420 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.473438 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.473449 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.502926 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:06 crc kubenswrapper[4799]: E0930 14:21:06.503043 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.503214 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:06 crc kubenswrapper[4799]: E0930 14:21:06.503286 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.575351 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.575404 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.575419 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.575433 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.575441 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.678109 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.678161 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.678174 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.678191 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.678203 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.780494 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.780540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.780555 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.780575 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.780588 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.882839 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.882867 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.882874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.882887 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.882897 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.985718 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.985759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.985772 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.985789 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:06 crc kubenswrapper[4799]: I0930 14:21:06.985800 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:06Z","lastTransitionTime":"2025-09-30T14:21:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.088168 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.088217 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.088230 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.088247 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.088262 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.191189 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.191228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.191237 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.191251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.191260 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.294221 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.294267 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.294278 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.294295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.294306 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.396922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.396960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.396968 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.396984 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.396994 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.499306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.499348 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.499361 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.499379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.499390 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.502842 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.502884 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:07 crc kubenswrapper[4799]: E0930 14:21:07.502947 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:07 crc kubenswrapper[4799]: E0930 14:21:07.503067 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.603679 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.603722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.603742 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.603761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.603776 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.707054 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.707126 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.707142 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.707171 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.707191 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.809540 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.809585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.809597 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.809615 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.809690 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.912833 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.912907 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.912932 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.912963 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:07 crc kubenswrapper[4799]: I0930 14:21:07.912984 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:07Z","lastTransitionTime":"2025-09-30T14:21:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.015585 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.015625 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.015634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.015667 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.015679 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.117898 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.117935 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.117946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.117962 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.117972 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.219929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.219965 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.219975 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.220039 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.220056 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.322418 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.322452 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.322462 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.322478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.322488 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.424304 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.424346 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.424356 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.424372 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.424382 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.503176 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.503324 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:08 crc kubenswrapper[4799]: E0930 14:21:08.503460 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:08 crc kubenswrapper[4799]: E0930 14:21:08.503486 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.525117 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:59Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-30T14:20:52Z\\\",\\\"message\\\":\\\"ft-image-registry/node-ca-vmvlc openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/machine-config-daemon-wr2vb openshift-multus/network-metrics-daemon-hghbr openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-etcd/etcd-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-additional-cni-plugins-hstq9]\\\\nI0930 14:20:52.378887 6714 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0930 14:20:52.378902 6714 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nI0930 14:20:52.378913 6714 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nI0930 14:20:52.378924 6714 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-hstq9 in node crc\\\\nI0930 14:20:52.378929 6714 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-hstq9 after 0 failed attempt(s)\\\\nI0930 14:20:52.378936 6714 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-hstq9\\\\nF0930 14:20:52.378952 6714 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gh9hh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6bhnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.529077 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.529120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.529136 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.529154 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.529165 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.537489 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmvlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06061f81-237c-4944-b7f2-0352d21ff125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cad095d04226451fb3bbdd8ddf5f236849025be20d3d29ab43c62fe6aa81a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5tgs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:20:01Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmvlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.556094 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7bd796c-f6f5-46c8-bf05-d43a170654ac\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da7db2d49de12fa1d36f8f3120cba985b8e14b815987ac112b48006c96247b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9f6a187c1b5611ec69b621f35e615945d98672723b28f79d7269115008e53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://071b644891c9e135b301319e4023cbb813f843c37d64efb04ccb44306ded91c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93bcc4b94660048ab01897c6e2b3b91cca31a5d3f98da628c18d873ece87883c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83c2146dc6f5b2deda800dfe78760dd77b3d15c62bbd2dbc1376c8ca9af51c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db48cbcabeebd136d713cc8e37c3c319552c364e4f545735a82db43c0e253f00\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de85794c03fd34b4b9112185d46868450fa5e476b54130a5d461ab8ae1f246be\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45674e4b619baec4e25aecb629d63ec85618a9750161606dd64b1513bbc5f835\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.568673 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e6f5c3b-51f0-4c8b-9032-0c122aedc898\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-30T14:19:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b48fb5859bdd7a18e39cb60e5571894a5ad6310891448a32991d504ecd7a34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e010d9520572bd2af7bbeb36b414012aeb2c7b20c00277d74a412405436771e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb174f5d070bebd94be20d1165385a970404cd05909b016a0ff028d2c978c52b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5e90de6d18dc88e899bb7dda86b936333757d67205c2bae6bd8ecf70fe26286\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-30T14:19:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-30T14:19:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-30T14:19:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.582228 4799 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-30T14:20:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143d95ad38cf347a48deb08709a43aa1a59e0ff6eb0bd9e2c6d74ff883f95e0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:20:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e83e1124f1e5af947ea94aa4934ea45fea9c505d1829b52c70f69cb016028a83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-30T14:19:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-30T14:21:08Z is after 2025-08-24T17:21:41Z" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.631826 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.631870 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.631880 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.631897 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.631909 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.667233 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podStartSLOduration=70.667209218 podStartE2EDuration="1m10.667209218s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:08.64298541 +0000 UTC m=+90.726585847" watchObservedRunningTime="2025-09-30 14:21:08.667209218 +0000 UTC m=+90.750809645" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.680965 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-hstq9" podStartSLOduration=70.680942995 podStartE2EDuration="1m10.680942995s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:08.667150197 +0000 UTC m=+90.750750634" watchObservedRunningTime="2025-09-30 14:21:08.680942995 +0000 UTC m=+90.764543422" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.714139 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-sjjm8" podStartSLOduration=70.714105169 podStartE2EDuration="1m10.714105169s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:08.713221318 +0000 UTC m=+90.796821745" watchObservedRunningTime="2025-09-30 14:21:08.714105169 +0000 UTC m=+90.797705596" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.725612 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hlg57" podStartSLOduration=70.725580833 podStartE2EDuration="1m10.725580833s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:08.72503046 +0000 UTC m=+90.808630887" watchObservedRunningTime="2025-09-30 14:21:08.725580833 +0000 UTC m=+90.809181260" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.734113 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.734139 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.734146 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.734175 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.734188 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.767116 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7vhf8" podStartSLOduration=69.76709711 podStartE2EDuration="1m9.76709711s" podCreationTimestamp="2025-09-30 14:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:08.767088809 +0000 UTC m=+90.850689236" watchObservedRunningTime="2025-09-30 14:21:08.76709711 +0000 UTC m=+90.850697527" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.796076 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.796060307 podStartE2EDuration="1m9.796060307s" podCreationTimestamp="2025-09-30 14:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:08.795660378 +0000 UTC m=+90.879260805" watchObservedRunningTime="2025-09-30 14:21:08.796060307 +0000 UTC m=+90.879660734" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.809837 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.809820804 podStartE2EDuration="1m9.809820804s" podCreationTimestamp="2025-09-30 14:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:08.809408314 +0000 UTC m=+90.893008741" watchObservedRunningTime="2025-09-30 14:21:08.809820804 +0000 UTC m=+90.893421241" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.823803 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=15.823784216 podStartE2EDuration="15.823784216s" podCreationTimestamp="2025-09-30 14:20:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:08.823477359 +0000 UTC m=+90.907077796" watchObservedRunningTime="2025-09-30 14:21:08.823784216 +0000 UTC m=+90.907384643" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.837148 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.837206 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.837216 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.837240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.837251 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.940888 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.940946 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.940960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.940983 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:08 crc kubenswrapper[4799]: I0930 14:21:08.940997 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:08Z","lastTransitionTime":"2025-09-30T14:21:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.043714 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.043775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.043789 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.043815 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.043829 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.146053 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.146093 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.146103 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.146119 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.146131 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.249824 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.249874 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.249886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.249903 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.249914 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.352882 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.352927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.352936 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.352954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.352965 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.456446 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.456507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.456522 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.456547 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.456563 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.503126 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.503225 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:09 crc kubenswrapper[4799]: E0930 14:21:09.503262 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:09 crc kubenswrapper[4799]: E0930 14:21:09.503397 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.559748 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.559803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.559817 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.559838 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.559850 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.662802 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.662886 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.662906 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.662939 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.662961 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.766740 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.766804 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.766820 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.766846 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.766860 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.869285 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.869317 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.869325 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.869354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.869363 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.971922 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.971952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.971960 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.971973 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:09 crc kubenswrapper[4799]: I0930 14:21:09.971982 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:09Z","lastTransitionTime":"2025-09-30T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.074532 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.074590 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.074609 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.074634 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.074651 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.177744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.177803 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.177816 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.177835 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.177846 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.280307 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.280366 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.280379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.280400 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.280686 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.384332 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.384469 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.384487 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.384507 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.384521 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.486917 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.486961 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.486972 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.486989 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.486999 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.502598 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.502614 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:10 crc kubenswrapper[4799]: E0930 14:21:10.502989 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:10 crc kubenswrapper[4799]: E0930 14:21:10.503056 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.588684 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.588774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.588786 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.588801 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.588815 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.691088 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.691150 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.691162 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.691180 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.691211 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.793793 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.793875 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.793888 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.793904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.793916 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.896643 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.896711 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.896726 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.896742 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:10 crc kubenswrapper[4799]: I0930 14:21:10.896755 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:10Z","lastTransitionTime":"2025-09-30T14:21:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.001240 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.001443 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.001533 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.001632 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.001729 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.104787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.104832 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.104843 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.104859 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.104876 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.208233 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.208304 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.208319 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.208345 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.208362 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.311295 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.311354 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.311375 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.311399 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.311412 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.415118 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.415212 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.415265 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.415306 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.415339 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.502217 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.502340 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:11 crc kubenswrapper[4799]: E0930 14:21:11.502396 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:11 crc kubenswrapper[4799]: E0930 14:21:11.502517 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.517949 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.517999 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.518008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.518024 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.518037 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.620447 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.620478 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.620488 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.620501 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.620509 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.722731 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.722771 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.722782 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.722817 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.722828 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.825382 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.825425 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.825434 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.825448 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.825459 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.927610 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.927698 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.927708 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.927724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:11 crc kubenswrapper[4799]: I0930 14:21:11.927733 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:11Z","lastTransitionTime":"2025-09-30T14:21:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.029712 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.029763 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.029774 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.029794 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.029808 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.133019 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.133066 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.133079 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.133096 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.133109 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.235503 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.235572 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.235587 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.235605 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.235629 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.337941 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.337985 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.338001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.338026 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.338040 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.440822 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.440889 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.440899 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.440915 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.440926 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.502956 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.503040 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:12 crc kubenswrapper[4799]: E0930 14:21:12.503544 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:12 crc kubenswrapper[4799]: E0930 14:21:12.503693 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.543930 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.543987 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.543997 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.544017 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.544030 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.647228 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.647589 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.647601 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.647655 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.647687 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.750967 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.751048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.751067 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.751120 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.751138 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.854248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.854301 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.854314 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.854334 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.854367 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.957333 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.957474 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.957486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.957505 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:12 crc kubenswrapper[4799]: I0930 14:21:12.957517 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:12Z","lastTransitionTime":"2025-09-30T14:21:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.060643 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.060723 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.060734 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.060756 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.060772 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.163617 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.163904 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.163970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.164036 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.164103 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.266243 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.266270 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.266280 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.266293 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.266302 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.369379 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.369407 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.369417 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.369431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.369441 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.471889 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.471915 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.471927 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.471942 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.471952 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.502918 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:13 crc kubenswrapper[4799]: E0930 14:21:13.503040 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.503740 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:13 crc kubenswrapper[4799]: E0930 14:21:13.503946 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.574716 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.574744 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.574754 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.574769 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.574781 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.677324 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.677360 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.677371 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.677388 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.677399 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.779828 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.779852 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.779860 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.779873 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.779882 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.881639 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.881710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.881722 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.881738 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.881749 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.985203 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.985251 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.985261 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.985277 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:13 crc kubenswrapper[4799]: I0930 14:21:13.985288 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:13Z","lastTransitionTime":"2025-09-30T14:21:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.088020 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.088061 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.088073 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.088088 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.088105 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.190715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.190759 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.190775 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.190790 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.190800 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.292885 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.292912 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.292920 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.292933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.292941 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.394943 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.394970 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.394978 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.394992 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.395003 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.497187 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.497224 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.497234 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.497248 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.497259 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.502472 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.502474 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:14 crc kubenswrapper[4799]: E0930 14:21:14.502747 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:14 crc kubenswrapper[4799]: E0930 14:21:14.502912 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.600404 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.600461 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.600471 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.600486 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.600495 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.703633 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.703700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.703715 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.703736 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.703748 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.807686 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.807761 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.807787 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.807814 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.807830 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.910929 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.910980 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.910992 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.911009 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:14 crc kubenswrapper[4799]: I0930 14:21:14.911021 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:14Z","lastTransitionTime":"2025-09-30T14:21:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.013866 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.013933 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.013952 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.013979 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.013998 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.116499 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.116535 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.116543 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.116557 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.116567 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.218974 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.219008 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.219018 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.219048 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.219064 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.321951 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.322001 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.322014 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.322040 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.322054 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.425269 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.425312 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.425320 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.425338 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.425349 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.503160 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.503160 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:15 crc kubenswrapper[4799]: E0930 14:21:15.503507 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:15 crc kubenswrapper[4799]: E0930 14:21:15.503632 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.527563 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.527602 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.527613 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.527631 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.527642 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.630260 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.630431 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.630470 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.630506 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.630533 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.734011 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.734059 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.734068 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.734091 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.734103 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.836414 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.836472 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.836483 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.836502 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.836514 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.938654 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.938700 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.938710 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.938724 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:15 crc kubenswrapper[4799]: I0930 14:21:15.938734 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:15Z","lastTransitionTime":"2025-09-30T14:21:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.041893 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.041944 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.041954 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.041971 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.041983 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:16Z","lastTransitionTime":"2025-09-30T14:21:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.145389 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.145444 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.145455 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.145476 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.145492 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:16Z","lastTransitionTime":"2025-09-30T14:21:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.248601 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.248685 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.248701 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.248727 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.248744 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:16Z","lastTransitionTime":"2025-09-30T14:21:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.348147 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.348205 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.348224 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.348246 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.348259 4799 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-30T14:21:16Z","lastTransitionTime":"2025-09-30T14:21:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.399974 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv"] Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.400385 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.403952 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.404404 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.404709 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.404912 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.458317 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-vmvlc" podStartSLOduration=78.458285782 podStartE2EDuration="1m18.458285782s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:16.442883997 +0000 UTC m=+98.526484414" watchObservedRunningTime="2025-09-30 14:21:16.458285782 +0000 UTC m=+98.541886209" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.470359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/100999a9-12c4-461d-bbf9-af7bc72e14fc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.470407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/100999a9-12c4-461d-bbf9-af7bc72e14fc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.470427 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/100999a9-12c4-461d-bbf9-af7bc72e14fc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.470486 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/100999a9-12c4-461d-bbf9-af7bc72e14fc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.470514 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/100999a9-12c4-461d-bbf9-af7bc72e14fc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.484342 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=75.484322102 podStartE2EDuration="1m15.484322102s" podCreationTimestamp="2025-09-30 14:20:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:16.48421092 +0000 UTC m=+98.567811357" watchObservedRunningTime="2025-09-30 14:21:16.484322102 +0000 UTC m=+98.567922519" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.499224 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.499203035 podStartE2EDuration="44.499203035s" podCreationTimestamp="2025-09-30 14:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:16.498193862 +0000 UTC m=+98.581794279" watchObservedRunningTime="2025-09-30 14:21:16.499203035 +0000 UTC m=+98.582803462" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.504175 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:16 crc kubenswrapper[4799]: E0930 14:21:16.504716 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.504174 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:16 crc kubenswrapper[4799]: E0930 14:21:16.505045 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.571729 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/100999a9-12c4-461d-bbf9-af7bc72e14fc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.571786 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/100999a9-12c4-461d-bbf9-af7bc72e14fc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.571792 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/100999a9-12c4-461d-bbf9-af7bc72e14fc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.571825 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/100999a9-12c4-461d-bbf9-af7bc72e14fc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.571858 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/100999a9-12c4-461d-bbf9-af7bc72e14fc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.571897 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/100999a9-12c4-461d-bbf9-af7bc72e14fc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.571955 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/100999a9-12c4-461d-bbf9-af7bc72e14fc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.571976 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:16 crc kubenswrapper[4799]: E0930 14:21:16.572053 4799 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:21:16 crc kubenswrapper[4799]: E0930 14:21:16.572102 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs podName:15ddd33f-0b7f-4cb5-b09a-03c211c759c8 nodeName:}" failed. No retries permitted until 2025-09-30 14:22:20.572088464 +0000 UTC m=+162.655688891 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs") pod "network-metrics-daemon-hghbr" (UID: "15ddd33f-0b7f-4cb5-b09a-03c211c759c8") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.572846 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/100999a9-12c4-461d-bbf9-af7bc72e14fc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.584142 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/100999a9-12c4-461d-bbf9-af7bc72e14fc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.591392 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/100999a9-12c4-461d-bbf9-af7bc72e14fc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-cnzxv\" (UID: \"100999a9-12c4-461d-bbf9-af7bc72e14fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:16 crc kubenswrapper[4799]: I0930 14:21:16.715134 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" Sep 30 14:21:17 crc kubenswrapper[4799]: I0930 14:21:17.023584 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" event={"ID":"100999a9-12c4-461d-bbf9-af7bc72e14fc","Type":"ContainerStarted","Data":"5371d4675c7c73290023ba8ce2f2691053b493a9cde8cf35e8cceeeaf9232ed0"} Sep 30 14:21:17 crc kubenswrapper[4799]: I0930 14:21:17.023639 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" event={"ID":"100999a9-12c4-461d-bbf9-af7bc72e14fc","Type":"ContainerStarted","Data":"9da4edbcc4d28362781cdb03c1f442d98dcdcab1162bb66016ba25feaf4d1bab"} Sep 30 14:21:17 crc kubenswrapper[4799]: I0930 14:21:17.042550 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cnzxv" podStartSLOduration=79.042530343 podStartE2EDuration="1m19.042530343s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:17.042196756 +0000 UTC m=+99.125797183" watchObservedRunningTime="2025-09-30 14:21:17.042530343 +0000 UTC m=+99.126130770" Sep 30 14:21:17 crc kubenswrapper[4799]: I0930 14:21:17.502417 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:17 crc kubenswrapper[4799]: I0930 14:21:17.502434 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:17 crc kubenswrapper[4799]: E0930 14:21:17.503088 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:17 crc kubenswrapper[4799]: E0930 14:21:17.503226 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:18 crc kubenswrapper[4799]: I0930 14:21:18.502204 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:18 crc kubenswrapper[4799]: E0930 14:21:18.502331 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:18 crc kubenswrapper[4799]: I0930 14:21:18.502430 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:18 crc kubenswrapper[4799]: E0930 14:21:18.502575 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:19 crc kubenswrapper[4799]: I0930 14:21:19.502359 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:19 crc kubenswrapper[4799]: I0930 14:21:19.502474 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:19 crc kubenswrapper[4799]: E0930 14:21:19.502478 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:19 crc kubenswrapper[4799]: E0930 14:21:19.502585 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:19 crc kubenswrapper[4799]: I0930 14:21:19.503174 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:21:19 crc kubenswrapper[4799]: E0930 14:21:19.503302 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6bhnh_openshift-ovn-kubernetes(fcc6c6df-d8b4-4d81-9aa5-8bc35533304b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" Sep 30 14:21:20 crc kubenswrapper[4799]: I0930 14:21:20.502969 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:20 crc kubenswrapper[4799]: I0930 14:21:20.503147 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:20 crc kubenswrapper[4799]: E0930 14:21:20.503290 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:20 crc kubenswrapper[4799]: E0930 14:21:20.503486 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:21 crc kubenswrapper[4799]: I0930 14:21:21.503175 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:21 crc kubenswrapper[4799]: I0930 14:21:21.503255 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:21 crc kubenswrapper[4799]: E0930 14:21:21.503312 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:21 crc kubenswrapper[4799]: E0930 14:21:21.503433 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:22 crc kubenswrapper[4799]: I0930 14:21:22.503108 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:22 crc kubenswrapper[4799]: I0930 14:21:22.503155 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:22 crc kubenswrapper[4799]: E0930 14:21:22.503263 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:22 crc kubenswrapper[4799]: E0930 14:21:22.503463 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:23 crc kubenswrapper[4799]: I0930 14:21:23.502826 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:23 crc kubenswrapper[4799]: I0930 14:21:23.502991 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:23 crc kubenswrapper[4799]: E0930 14:21:23.503133 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:23 crc kubenswrapper[4799]: E0930 14:21:23.503371 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:24 crc kubenswrapper[4799]: I0930 14:21:24.502702 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:24 crc kubenswrapper[4799]: I0930 14:21:24.502744 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:24 crc kubenswrapper[4799]: E0930 14:21:24.502850 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:24 crc kubenswrapper[4799]: E0930 14:21:24.502933 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:25 crc kubenswrapper[4799]: I0930 14:21:25.503024 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:25 crc kubenswrapper[4799]: I0930 14:21:25.503121 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:25 crc kubenswrapper[4799]: E0930 14:21:25.503260 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:25 crc kubenswrapper[4799]: E0930 14:21:25.503400 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:26 crc kubenswrapper[4799]: I0930 14:21:26.502097 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:26 crc kubenswrapper[4799]: I0930 14:21:26.502097 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:26 crc kubenswrapper[4799]: E0930 14:21:26.502272 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:26 crc kubenswrapper[4799]: E0930 14:21:26.502386 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:27 crc kubenswrapper[4799]: I0930 14:21:27.502463 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:27 crc kubenswrapper[4799]: I0930 14:21:27.503034 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:27 crc kubenswrapper[4799]: E0930 14:21:27.503681 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:27 crc kubenswrapper[4799]: E0930 14:21:27.503836 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:28 crc kubenswrapper[4799]: I0930 14:21:28.502852 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:28 crc kubenswrapper[4799]: E0930 14:21:28.504586 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:28 crc kubenswrapper[4799]: I0930 14:21:28.504609 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:28 crc kubenswrapper[4799]: E0930 14:21:28.504899 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:29 crc kubenswrapper[4799]: I0930 14:21:29.502689 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:29 crc kubenswrapper[4799]: I0930 14:21:29.502735 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:29 crc kubenswrapper[4799]: E0930 14:21:29.502820 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:29 crc kubenswrapper[4799]: E0930 14:21:29.502977 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:30 crc kubenswrapper[4799]: I0930 14:21:30.502837 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:30 crc kubenswrapper[4799]: E0930 14:21:30.503031 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:30 crc kubenswrapper[4799]: I0930 14:21:30.503417 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:30 crc kubenswrapper[4799]: E0930 14:21:30.503529 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:31 crc kubenswrapper[4799]: I0930 14:21:31.502914 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:31 crc kubenswrapper[4799]: I0930 14:21:31.502947 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:31 crc kubenswrapper[4799]: E0930 14:21:31.503115 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:31 crc kubenswrapper[4799]: E0930 14:21:31.503253 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:32 crc kubenswrapper[4799]: I0930 14:21:32.502731 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:32 crc kubenswrapper[4799]: E0930 14:21:32.502970 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:32 crc kubenswrapper[4799]: I0930 14:21:32.503454 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:32 crc kubenswrapper[4799]: E0930 14:21:32.503575 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:33 crc kubenswrapper[4799]: I0930 14:21:33.502489 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:33 crc kubenswrapper[4799]: E0930 14:21:33.502862 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:33 crc kubenswrapper[4799]: I0930 14:21:33.502565 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:33 crc kubenswrapper[4799]: E0930 14:21:33.503485 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.080191 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/1.log" Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.080688 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/0.log" Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.080779 4799 generic.go:334] "Generic (PLEG): container finished" podID="5517db4f-9237-4b92-a672-6fa7afcd7e3f" containerID="f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb" exitCode=1 Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.080819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sjjm8" event={"ID":"5517db4f-9237-4b92-a672-6fa7afcd7e3f","Type":"ContainerDied","Data":"f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb"} Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.080857 4799 scope.go:117] "RemoveContainer" containerID="f8432587a7df2f134537a083ec497ce4d1b05291714053d540ef41f8998b89de" Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.081499 4799 scope.go:117] "RemoveContainer" containerID="f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb" Sep 30 14:21:34 crc kubenswrapper[4799]: E0930 14:21:34.081761 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-sjjm8_openshift-multus(5517db4f-9237-4b92-a672-6fa7afcd7e3f)\"" pod="openshift-multus/multus-sjjm8" podUID="5517db4f-9237-4b92-a672-6fa7afcd7e3f" Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.502457 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:34 crc kubenswrapper[4799]: E0930 14:21:34.502590 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.502457 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:34 crc kubenswrapper[4799]: E0930 14:21:34.503053 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:34 crc kubenswrapper[4799]: I0930 14:21:34.503269 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.086980 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/1.log" Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.091192 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/3.log" Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.093716 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerStarted","Data":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.095093 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.122369 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podStartSLOduration=97.122231668 podStartE2EDuration="1m37.122231668s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:21:35.121409699 +0000 UTC m=+117.205010146" watchObservedRunningTime="2025-09-30 14:21:35.122231668 +0000 UTC m=+117.205832095" Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.284967 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hghbr"] Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.285158 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:35 crc kubenswrapper[4799]: E0930 14:21:35.285280 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.502230 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:35 crc kubenswrapper[4799]: I0930 14:21:35.502262 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:35 crc kubenswrapper[4799]: E0930 14:21:35.502338 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:35 crc kubenswrapper[4799]: E0930 14:21:35.502511 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:36 crc kubenswrapper[4799]: I0930 14:21:36.502526 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:36 crc kubenswrapper[4799]: E0930 14:21:36.502746 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:37 crc kubenswrapper[4799]: I0930 14:21:37.502759 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:37 crc kubenswrapper[4799]: I0930 14:21:37.502769 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:37 crc kubenswrapper[4799]: E0930 14:21:37.502974 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:37 crc kubenswrapper[4799]: E0930 14:21:37.503071 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:37 crc kubenswrapper[4799]: I0930 14:21:37.502797 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:37 crc kubenswrapper[4799]: E0930 14:21:37.503246 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:38 crc kubenswrapper[4799]: E0930 14:21:38.459702 4799 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Sep 30 14:21:38 crc kubenswrapper[4799]: I0930 14:21:38.503058 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:38 crc kubenswrapper[4799]: E0930 14:21:38.505847 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:38 crc kubenswrapper[4799]: E0930 14:21:38.590015 4799 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 14:21:39 crc kubenswrapper[4799]: I0930 14:21:39.503088 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:39 crc kubenswrapper[4799]: I0930 14:21:39.503117 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:39 crc kubenswrapper[4799]: I0930 14:21:39.503134 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:39 crc kubenswrapper[4799]: E0930 14:21:39.503256 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:39 crc kubenswrapper[4799]: E0930 14:21:39.503357 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:39 crc kubenswrapper[4799]: E0930 14:21:39.503487 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:40 crc kubenswrapper[4799]: I0930 14:21:40.502818 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:40 crc kubenswrapper[4799]: E0930 14:21:40.502962 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:41 crc kubenswrapper[4799]: I0930 14:21:41.503202 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:41 crc kubenswrapper[4799]: E0930 14:21:41.504271 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:41 crc kubenswrapper[4799]: I0930 14:21:41.503260 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:41 crc kubenswrapper[4799]: E0930 14:21:41.504486 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:41 crc kubenswrapper[4799]: I0930 14:21:41.503229 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:41 crc kubenswrapper[4799]: E0930 14:21:41.504693 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:42 crc kubenswrapper[4799]: I0930 14:21:42.503306 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:42 crc kubenswrapper[4799]: E0930 14:21:42.503541 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:43 crc kubenswrapper[4799]: I0930 14:21:43.502329 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:43 crc kubenswrapper[4799]: I0930 14:21:43.502373 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:43 crc kubenswrapper[4799]: I0930 14:21:43.502398 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:43 crc kubenswrapper[4799]: E0930 14:21:43.502461 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:43 crc kubenswrapper[4799]: E0930 14:21:43.502508 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:43 crc kubenswrapper[4799]: E0930 14:21:43.502554 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:43 crc kubenswrapper[4799]: E0930 14:21:43.591100 4799 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 14:21:44 crc kubenswrapper[4799]: I0930 14:21:44.503157 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:44 crc kubenswrapper[4799]: E0930 14:21:44.503298 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:45 crc kubenswrapper[4799]: I0930 14:21:45.502202 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:45 crc kubenswrapper[4799]: E0930 14:21:45.502547 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:45 crc kubenswrapper[4799]: I0930 14:21:45.503096 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:45 crc kubenswrapper[4799]: E0930 14:21:45.503273 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:45 crc kubenswrapper[4799]: I0930 14:21:45.503628 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:45 crc kubenswrapper[4799]: E0930 14:21:45.503846 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:46 crc kubenswrapper[4799]: I0930 14:21:46.502762 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:46 crc kubenswrapper[4799]: E0930 14:21:46.503455 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:47 crc kubenswrapper[4799]: I0930 14:21:47.502933 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:47 crc kubenswrapper[4799]: I0930 14:21:47.502963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:47 crc kubenswrapper[4799]: I0930 14:21:47.502962 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:47 crc kubenswrapper[4799]: E0930 14:21:47.503070 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:47 crc kubenswrapper[4799]: E0930 14:21:47.503132 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:47 crc kubenswrapper[4799]: E0930 14:21:47.503205 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:48 crc kubenswrapper[4799]: I0930 14:21:48.502309 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:48 crc kubenswrapper[4799]: E0930 14:21:48.503285 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:48 crc kubenswrapper[4799]: I0930 14:21:48.503910 4799 scope.go:117] "RemoveContainer" containerID="f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb" Sep 30 14:21:48 crc kubenswrapper[4799]: E0930 14:21:48.593257 4799 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 30 14:21:49 crc kubenswrapper[4799]: I0930 14:21:49.137187 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/1.log" Sep 30 14:21:49 crc kubenswrapper[4799]: I0930 14:21:49.137502 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sjjm8" event={"ID":"5517db4f-9237-4b92-a672-6fa7afcd7e3f","Type":"ContainerStarted","Data":"52713cf3a77276c7a876e200beda5ecf6f3f1a5e784f0f9c2bb82d70e6d3f2b9"} Sep 30 14:21:49 crc kubenswrapper[4799]: I0930 14:21:49.502930 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:49 crc kubenswrapper[4799]: I0930 14:21:49.502982 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:49 crc kubenswrapper[4799]: I0930 14:21:49.503024 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:49 crc kubenswrapper[4799]: E0930 14:21:49.503071 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:49 crc kubenswrapper[4799]: E0930 14:21:49.503159 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:49 crc kubenswrapper[4799]: E0930 14:21:49.503232 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:50 crc kubenswrapper[4799]: I0930 14:21:50.502929 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:50 crc kubenswrapper[4799]: E0930 14:21:50.503064 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:51 crc kubenswrapper[4799]: I0930 14:21:51.503157 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:51 crc kubenswrapper[4799]: I0930 14:21:51.503210 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:51 crc kubenswrapper[4799]: I0930 14:21:51.503215 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:51 crc kubenswrapper[4799]: E0930 14:21:51.503300 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:51 crc kubenswrapper[4799]: E0930 14:21:51.503510 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:51 crc kubenswrapper[4799]: E0930 14:21:51.503685 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:52 crc kubenswrapper[4799]: I0930 14:21:52.502989 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:52 crc kubenswrapper[4799]: E0930 14:21:52.503144 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 30 14:21:53 crc kubenswrapper[4799]: I0930 14:21:53.502926 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:53 crc kubenswrapper[4799]: I0930 14:21:53.502967 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:53 crc kubenswrapper[4799]: I0930 14:21:53.502945 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:53 crc kubenswrapper[4799]: E0930 14:21:53.503091 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 30 14:21:53 crc kubenswrapper[4799]: E0930 14:21:53.503145 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hghbr" podUID="15ddd33f-0b7f-4cb5-b09a-03c211c759c8" Sep 30 14:21:53 crc kubenswrapper[4799]: E0930 14:21:53.503250 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 30 14:21:54 crc kubenswrapper[4799]: I0930 14:21:54.503177 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:21:54 crc kubenswrapper[4799]: I0930 14:21:54.505827 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 30 14:21:54 crc kubenswrapper[4799]: I0930 14:21:54.506047 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 30 14:21:55 crc kubenswrapper[4799]: I0930 14:21:55.502855 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:21:55 crc kubenswrapper[4799]: I0930 14:21:55.502855 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:21:55 crc kubenswrapper[4799]: I0930 14:21:55.502861 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:21:55 crc kubenswrapper[4799]: I0930 14:21:55.505394 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Sep 30 14:21:55 crc kubenswrapper[4799]: I0930 14:21:55.505487 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Sep 30 14:21:55 crc kubenswrapper[4799]: I0930 14:21:55.516446 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 30 14:21:55 crc kubenswrapper[4799]: I0930 14:21:55.518698 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.262387 4799 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.302756 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xvd7z"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.303362 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.306737 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.306796 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.307098 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.308144 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.308673 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.308849 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.309514 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.309862 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-clzwz"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.310481 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.310640 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.312199 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.314117 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.314783 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.316108 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.316476 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.317149 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.317932 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.331877 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.333816 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.334105 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.338174 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.353271 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.353721 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfxv2"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.354031 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmw7h"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.354385 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jgd4t"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.354479 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.354813 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.355022 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.355286 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.355598 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.357495 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.357714 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.357939 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.358080 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.358234 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.359006 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.359522 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.360188 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.360381 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.360519 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.360641 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.360769 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.360885 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.361090 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.365971 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.367021 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.367756 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.368261 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.368763 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.369543 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w9q9n"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.369879 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.370096 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.370440 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.372287 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-7f9sl"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.372916 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.373406 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.373843 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.375256 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.375389 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.383206 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.383462 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.383812 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.384169 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.384898 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.385051 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.385318 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.385623 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.385813 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.385925 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.385972 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386097 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386187 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386249 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386367 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386421 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386518 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386585 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386724 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.386727 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dzk4q"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.387131 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.387534 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-svk57"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.387959 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.388255 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.388366 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pnkgs"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.389133 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.394926 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.395215 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-clzwz"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.401515 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.401780 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.401842 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.402738 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.402787 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.403435 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.406217 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411288 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-etcd-client\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411330 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-trusted-ca\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk87g\" (UniqueName: \"kubernetes.io/projected/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-kube-api-access-vk87g\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-client-ca\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411516 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvtjl\" (UniqueName: \"kubernetes.io/projected/dedc044b-a09f-436a-8bdb-3b606269e3c5-kube-api-access-lvtjl\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411586 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-serving-cert\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411694 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411718 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dedc044b-a09f-436a-8bdb-3b606269e3c5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411754 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411775 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411838 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411859 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc8pm\" (UniqueName: \"kubernetes.io/projected/c35fadda-30b5-438e-b0a6-945004f36333-kube-api-access-jc8pm\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411902 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c35fadda-30b5-438e-b0a6-945004f36333-audit-dir\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411923 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/018cb267-3656-4202-ba7e-323811533a19-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411945 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.411980 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cskjn\" (UniqueName: \"kubernetes.io/projected/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-kube-api-access-cskjn\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412006 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412036 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412084 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-config\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412102 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fabab36-ee1e-4d92-b99c-b07ae87891be-serving-cert\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412135 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/018cb267-3656-4202-ba7e-323811533a19-config\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412157 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-config\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412174 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3981843-fd55-41a8-8377-a741fcd231fb-serving-cert\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412209 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-trusted-ca-bundle\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412235 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412255 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8209103b-69e0-41be-9ce9-c87dd3e01a19-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-plkz9\" (UID: \"8209103b-69e0-41be-9ce9-c87dd3e01a19\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412300 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66k57\" (UniqueName: \"kubernetes.io/projected/a3981843-fd55-41a8-8377-a741fcd231fb-kube-api-access-66k57\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412375 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqs2c\" (UniqueName: \"kubernetes.io/projected/9fabab36-ee1e-4d92-b99c-b07ae87891be-kube-api-access-lqs2c\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412410 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/018cb267-3656-4202-ba7e-323811533a19-images\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412445 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhnfk\" (UniqueName: \"kubernetes.io/projected/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-kube-api-access-vhnfk\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412463 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-serving-cert\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412527 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dedc044b-a09f-436a-8bdb-3b606269e3c5-serving-cert\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412546 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-config\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412565 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412600 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4pgb\" (UniqueName: \"kubernetes.io/projected/018cb267-3656-4202-ba7e-323811533a19-kube-api-access-l4pgb\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412619 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-node-pullsecrets\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412633 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412673 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412688 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-config\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-client-ca\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412723 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8bd5\" (UniqueName: \"kubernetes.io/projected/8209103b-69e0-41be-9ce9-c87dd3e01a19-kube-api-access-s8bd5\") pod \"cluster-samples-operator-665b6dd947-plkz9\" (UID: \"8209103b-69e0-41be-9ce9-c87dd3e01a19\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412759 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-audit-dir\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412778 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-audit\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412794 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-etcd-serving-ca\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412833 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-encryption-config\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412859 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-audit-policies\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.412879 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-image-import-ca\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.417991 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2tlvs"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.432575 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.434747 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.449259 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.449598 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.451934 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.455382 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.456135 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.456560 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.458012 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.458254 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.458426 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.461685 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.463208 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.468454 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.468765 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.469643 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.469746 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.478036 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.478211 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.478270 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.478306 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.478413 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.493432 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.493960 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.494346 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.495129 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.495212 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.496384 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.496627 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.497168 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.497286 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.499578 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.500722 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.500885 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.501125 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.501246 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.501336 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.501504 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.504961 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.505244 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.505438 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.505760 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.506050 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.506087 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.506179 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.506290 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.506686 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.507415 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.507866 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.508534 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.509631 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-glzch"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.508604 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.508642 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.508744 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.508771 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.513044 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.513818 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-config\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.518420 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fabab36-ee1e-4d92-b99c-b07ae87891be-serving-cert\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.518544 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3981843-fd55-41a8-8377-a741fcd231fb-serving-cert\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.518677 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c162443e-089d-49e2-968d-c712bf5f5aba-srv-cert\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.518774 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/018cb267-3656-4202-ba7e-323811533a19-config\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.518883 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-config\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.518968 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-encryption-config\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519043 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqd7n\" (UniqueName: \"kubernetes.io/projected/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-kube-api-access-wqd7n\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d01e81a-5012-4d30-863b-49396073cbd1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519189 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519269 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-trusted-ca-bundle\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-etcd-client\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519414 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8918fed-2347-4360-a4e1-d84658a2211f-audit-dir\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519491 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-machine-approver-tls\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519558 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519621 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8209103b-69e0-41be-9ce9-c87dd3e01a19-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-plkz9\" (UID: \"8209103b-69e0-41be-9ce9-c87dd3e01a19\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519730 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66k57\" (UniqueName: \"kubernetes.io/projected/a3981843-fd55-41a8-8377-a741fcd231fb-kube-api-access-66k57\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519838 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqs2c\" (UniqueName: \"kubernetes.io/projected/9fabab36-ee1e-4d92-b99c-b07ae87891be-kube-api-access-lqs2c\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.519935 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/018cb267-3656-4202-ba7e-323811533a19-images\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.520002 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-serving-cert\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.520069 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhnfk\" (UniqueName: \"kubernetes.io/projected/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-kube-api-access-vhnfk\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.520150 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.520258 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dedc044b-a09f-436a-8bdb-3b606269e3c5-serving-cert\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.520349 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-config\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.525793 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.525968 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-serving-cert\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526042 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-audit-policies\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526117 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s649c\" (UniqueName: \"kubernetes.io/projected/e2e2061e-3d24-41dd-b8ee-6b50e8313647-kube-api-access-s649c\") pod \"downloads-7954f5f757-7f9sl\" (UID: \"e2e2061e-3d24-41dd-b8ee-6b50e8313647\") " pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526190 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4pgb\" (UniqueName: \"kubernetes.io/projected/018cb267-3656-4202-ba7e-323811533a19-kube-api-access-l4pgb\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526263 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-node-pullsecrets\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526327 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-config\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526470 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-config\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526583 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-client-ca\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526686 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-auth-proxy-config\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526802 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526884 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-audit-dir\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.526952 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8bd5\" (UniqueName: \"kubernetes.io/projected/8209103b-69e0-41be-9ce9-c87dd3e01a19-kube-api-access-s8bd5\") pod \"cluster-samples-operator-665b6dd947-plkz9\" (UID: \"8209103b-69e0-41be-9ce9-c87dd3e01a19\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527027 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-audit\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527088 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-etcd-serving-ca\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527163 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-encryption-config\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527239 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-audit-policies\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-image-import-ca\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527391 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km7k7\" (UniqueName: \"kubernetes.io/projected/d8918fed-2347-4360-a4e1-d84658a2211f-kube-api-access-km7k7\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527560 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tml2q\" (UniqueName: \"kubernetes.io/projected/c162443e-089d-49e2-968d-c712bf5f5aba-kube-api-access-tml2q\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527678 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d01e81a-5012-4d30-863b-49396073cbd1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527771 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-etcd-client\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527861 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-trusted-ca\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.527941 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/383dabba-8b1c-4e84-945a-6d6b3749802c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528013 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c162443e-089d-49e2-968d-c712bf5f5aba-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528084 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528155 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d01e81a-5012-4d30-863b-49396073cbd1-config\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528230 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-client-ca\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528380 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk87g\" (UniqueName: \"kubernetes.io/projected/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-kube-api-access-vk87g\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvtjl\" (UniqueName: \"kubernetes.io/projected/dedc044b-a09f-436a-8bdb-3b606269e3c5-kube-api-access-lvtjl\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528558 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-serving-cert\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.528630 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/383dabba-8b1c-4e84-945a-6d6b3749802c-trusted-ca\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539033 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539113 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dedc044b-a09f-436a-8bdb-3b606269e3c5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539153 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98s4p\" (UniqueName: \"kubernetes.io/projected/383dabba-8b1c-4e84-945a-6d6b3749802c-kube-api-access-98s4p\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539178 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539230 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539256 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539320 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539347 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc8pm\" (UniqueName: \"kubernetes.io/projected/c35fadda-30b5-438e-b0a6-945004f36333-kube-api-access-jc8pm\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539369 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sv4n\" (UniqueName: \"kubernetes.io/projected/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-kube-api-access-8sv4n\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539415 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cskjn\" (UniqueName: \"kubernetes.io/projected/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-kube-api-access-cskjn\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539451 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c35fadda-30b5-438e-b0a6-945004f36333-audit-dir\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/018cb267-3656-4202-ba7e-323811533a19-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539493 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/383dabba-8b1c-4e84-945a-6d6b3749802c-metrics-tls\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539517 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539539 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.539563 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.543510 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/018cb267-3656-4202-ba7e-323811533a19-images\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.551831 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.552725 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.585263 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-trusted-ca-bundle\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.515329 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.587443 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.516185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-config\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.523028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-config\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.524972 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/018cb267-3656-4202-ba7e-323811533a19-config\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.587815 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-node-pullsecrets\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.588473 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-config\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.588663 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.589753 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.590022 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-config\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.591081 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-client-ca\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.591674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.591864 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-audit-dir\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.592664 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-audit\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.607131 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.607244 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8wmvv"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.607401 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-etcd-serving-ca\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.625979 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.628110 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.628692 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c35fadda-30b5-438e-b0a6-945004f36333-audit-dir\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.631223 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.637551 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.638007 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dedc044b-a09f-436a-8bdb-3b606269e3c5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.639342 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8209103b-69e0-41be-9ce9-c87dd3e01a19-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-plkz9\" (UID: \"8209103b-69e0-41be-9ce9-c87dd3e01a19\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.640496 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-audit-policies\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.641304 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-image-import-ca\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.644328 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/383dabba-8b1c-4e84-945a-6d6b3749802c-trusted-ca\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.647479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98s4p\" (UniqueName: \"kubernetes.io/projected/383dabba-8b1c-4e84-945a-6d6b3749802c-kube-api-access-98s4p\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.647559 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.647723 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sv4n\" (UniqueName: \"kubernetes.io/projected/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-kube-api-access-8sv4n\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.647894 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648007 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c162443e-089d-49e2-968d-c712bf5f5aba-srv-cert\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648085 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-encryption-config\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648157 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d01e81a-5012-4d30-863b-49396073cbd1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648235 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqd7n\" (UniqueName: \"kubernetes.io/projected/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-kube-api-access-wqd7n\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648325 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-etcd-client\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648405 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8918fed-2347-4360-a4e1-d84658a2211f-audit-dir\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648476 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-machine-approver-tls\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648622 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-serving-cert\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648693 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3981843-fd55-41a8-8377-a741fcd231fb-serving-cert\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648792 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-audit-policies\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648890 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s649c\" (UniqueName: \"kubernetes.io/projected/e2e2061e-3d24-41dd-b8ee-6b50e8313647-kube-api-access-s649c\") pod \"downloads-7954f5f757-7f9sl\" (UID: \"e2e2061e-3d24-41dd-b8ee-6b50e8313647\") " pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.648983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-config\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.649065 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-auth-proxy-config\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.649179 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km7k7\" (UniqueName: \"kubernetes.io/projected/d8918fed-2347-4360-a4e1-d84658a2211f-kube-api-access-km7k7\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.649260 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tml2q\" (UniqueName: \"kubernetes.io/projected/c162443e-089d-49e2-968d-c712bf5f5aba-kube-api-access-tml2q\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.649341 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d01e81a-5012-4d30-863b-49396073cbd1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.649419 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.649520 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.649637 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.649847 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.650163 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-serving-cert\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.650417 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dedc044b-a09f-436a-8bdb-3b606269e3c5-serving-cert\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.650561 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.650904 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-encryption-config\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.651154 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.646403 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fabab36-ee1e-4d92-b99c-b07ae87891be-serving-cert\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.645587 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/383dabba-8b1c-4e84-945a-6d6b3749802c-trusted-ca\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.516312 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.516491 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.520873 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.520937 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.541441 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.541564 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.553956 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.556180 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.557542 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.566911 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.665145 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-client-ca\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.666739 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.667188 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-etcd-client\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.668069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.668848 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.672203 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.675586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/018cb267-3656-4202-ba7e-323811533a19-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.689038 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.689426 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.690945 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-trusted-ca\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.691585 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-hrm8v"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.692076 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.692456 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.693125 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.693420 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.694439 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-auth-proxy-config\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.692455 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.697735 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.697786 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.698004 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.692604 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.695764 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.695805 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.699011 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.699637 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-audit-policies\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.701531 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.701678 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.706195 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-b7pv9"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.710375 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-machine-approver-tls\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.701804 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.702215 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8918fed-2347-4360-a4e1-d84658a2211f-audit-dir\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.709997 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.702641 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-config\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.711150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-encryption-config\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.711425 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.712260 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.712286 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.713242 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.713250 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.713664 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-serving-cert\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.714514 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.714565 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.717130 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmw7h"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.718065 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.718458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-serving-cert\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.719051 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jznkv"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.720002 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jznkv" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.721790 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dzk4q"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.723588 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.725269 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.725775 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.725776 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8918fed-2347-4360-a4e1-d84658a2211f-etcd-client\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.729021 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.729946 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jgd4t"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.731673 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-svk57"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.733097 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.734266 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfxv2"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.736074 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-m4jz9"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.736911 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.737679 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mkz8h"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.739640 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2tlvs"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.739730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.742281 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.743322 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w9q9n"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.745247 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.745712 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.746790 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.747109 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xvd7z"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.751694 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.751754 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/383dabba-8b1c-4e84-945a-6d6b3749802c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.751786 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c162443e-089d-49e2-968d-c712bf5f5aba-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.751809 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d01e81a-5012-4d30-863b-49396073cbd1-config\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.751911 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/383dabba-8b1c-4e84-945a-6d6b3749802c-metrics-tls\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.752676 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8918fed-2347-4360-a4e1-d84658a2211f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.753475 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.754023 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c162443e-089d-49e2-968d-c712bf5f5aba-srv-cert\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.754952 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pnkgs"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.756132 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/383dabba-8b1c-4e84-945a-6d6b3749802c-metrics-tls\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.756921 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8wmvv"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.759226 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jznkv"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.760147 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.762460 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7f9sl"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.765404 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.765688 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.768686 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.771638 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.787869 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.792900 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c162443e-089d-49e2-968d-c712bf5f5aba-profile-collector-cert\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.795287 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.797113 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-b7pv9"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.798636 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-hrm8v"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.800143 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.800568 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.801685 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.803538 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.804951 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mkz8h"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.807209 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.808776 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.810708 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vxm25"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.812204 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vxm25" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.813082 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.814639 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.817195 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vxm25"] Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.825682 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.845750 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.865862 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.885514 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.906484 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.925865 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.946741 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.966586 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.979489 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d01e81a-5012-4d30-863b-49396073cbd1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:57 crc kubenswrapper[4799]: I0930 14:21:57.986382 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.005831 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.026754 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.047352 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.066288 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.085779 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.106892 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.126234 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.134052 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d01e81a-5012-4d30-863b-49396073cbd1-config\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.146303 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.186179 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.226745 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.245854 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.266712 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.287491 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.305344 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.326203 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.346684 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.366378 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.386462 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.406012 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.427688 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.459165 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.465860 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.485222 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.506214 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.543777 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66k57\" (UniqueName: \"kubernetes.io/projected/a3981843-fd55-41a8-8377-a741fcd231fb-kube-api-access-66k57\") pod \"controller-manager-879f6c89f-rmw7h\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.543881 4799 request.go:700] Waited for 1.001174378s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.564363 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqs2c\" (UniqueName: \"kubernetes.io/projected/9fabab36-ee1e-4d92-b99c-b07ae87891be-kube-api-access-lqs2c\") pod \"route-controller-manager-6576b87f9c-xkv5k\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.567128 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.579420 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.586256 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.623905 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhnfk\" (UniqueName: \"kubernetes.io/projected/7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea-kube-api-access-vhnfk\") pod \"apiserver-76f77b778f-clzwz\" (UID: \"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea\") " pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.640721 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4pgb\" (UniqueName: \"kubernetes.io/projected/018cb267-3656-4202-ba7e-323811533a19-kube-api-access-l4pgb\") pod \"machine-api-operator-5694c8668f-jgd4t\" (UID: \"018cb267-3656-4202-ba7e-323811533a19\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.662472 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8bd5\" (UniqueName: \"kubernetes.io/projected/8209103b-69e0-41be-9ce9-c87dd3e01a19-kube-api-access-s8bd5\") pod \"cluster-samples-operator-665b6dd947-plkz9\" (UID: \"8209103b-69e0-41be-9ce9-c87dd3e01a19\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.662821 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.665530 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.702609 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cskjn\" (UniqueName: \"kubernetes.io/projected/3076dcdb-8618-4518-bf4a-bcf6b023cf0c-kube-api-access-cskjn\") pod \"openshift-apiserver-operator-796bbdcf4f-g4fl4\" (UID: \"3076dcdb-8618-4518-bf4a-bcf6b023cf0c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.711208 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.713975 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.720105 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc8pm\" (UniqueName: \"kubernetes.io/projected/c35fadda-30b5-438e-b0a6-945004f36333-kube-api-access-jc8pm\") pod \"oauth-openshift-558db77b4-bfxv2\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.725722 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.745846 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.788034 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.791481 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s649c\" (UniqueName: \"kubernetes.io/projected/e2e2061e-3d24-41dd-b8ee-6b50e8313647-kube-api-access-s649c\") pod \"downloads-7954f5f757-7f9sl\" (UID: \"e2e2061e-3d24-41dd-b8ee-6b50e8313647\") " pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.826429 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98s4p\" (UniqueName: \"kubernetes.io/projected/383dabba-8b1c-4e84-945a-6d6b3749802c-kube-api-access-98s4p\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.845176 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqd7n\" (UniqueName: \"kubernetes.io/projected/0095d26a-f43b-4d74-a34e-f5a3bf289dfd-kube-api-access-wqd7n\") pod \"machine-approver-56656f9798-6hvgg\" (UID: \"0095d26a-f43b-4d74-a34e-f5a3bf289dfd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.850562 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k"] Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.850823 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.861984 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" Sep 30 14:21:58 crc kubenswrapper[4799]: W0930 14:21:58.863119 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fabab36_ee1e_4d92_b99c_b07ae87891be.slice/crio-0a9b4bcbc351a7a5c77d7f232fb662559f925515ba44f307563e9a0b6ec04324 WatchSource:0}: Error finding container 0a9b4bcbc351a7a5c77d7f232fb662559f925515ba44f307563e9a0b6ec04324: Status 404 returned error can't find the container with id 0a9b4bcbc351a7a5c77d7f232fb662559f925515ba44f307563e9a0b6ec04324 Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.866923 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.885428 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.897146 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.897732 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmw7h"] Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.905710 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.906835 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.926485 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.934447 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.946307 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.947193 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jgd4t"] Sep 30 14:21:58 crc kubenswrapper[4799]: W0930 14:21:58.956861 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3981843_fd55_41a8_8377_a741fcd231fb.slice/crio-21544cd61b5b34562039e3b1fee260249bcd0faf29482405de90e86ac9d286f2 WatchSource:0}: Error finding container 21544cd61b5b34562039e3b1fee260249bcd0faf29482405de90e86ac9d286f2: Status 404 returned error can't find the container with id 21544cd61b5b34562039e3b1fee260249bcd0faf29482405de90e86ac9d286f2 Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.967463 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.973724 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:21:58 crc kubenswrapper[4799]: I0930 14:21:58.987055 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.006232 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.013359 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4"] Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.027271 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.048444 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.069841 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.086295 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.106449 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9"] Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.132152 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km7k7\" (UniqueName: \"kubernetes.io/projected/d8918fed-2347-4360-a4e1-d84658a2211f-kube-api-access-km7k7\") pod \"apiserver-7bbb656c7d-x52bz\" (UID: \"d8918fed-2347-4360-a4e1-d84658a2211f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.152616 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sv4n\" (UniqueName: \"kubernetes.io/projected/a2212e6e-a106-4efc-bb26-27bb27f8c0b8-kube-api-access-8sv4n\") pod \"openshift-controller-manager-operator-756b6f6bc6-p67w5\" (UID: \"a2212e6e-a106-4efc-bb26-27bb27f8c0b8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.160246 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d01e81a-5012-4d30-863b-49396073cbd1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xqj7j\" (UID: \"3d01e81a-5012-4d30-863b-49396073cbd1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.185555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvtjl\" (UniqueName: \"kubernetes.io/projected/dedc044b-a09f-436a-8bdb-3b606269e3c5-kube-api-access-lvtjl\") pod \"openshift-config-operator-7777fb866f-mgzkz\" (UID: \"dedc044b-a09f-436a-8bdb-3b606269e3c5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.202156 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfxv2"] Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.208007 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.218627 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk87g\" (UniqueName: \"kubernetes.io/projected/56b3121f-1c64-40d2-9cfd-b2c4a55451cc-kube-api-access-vk87g\") pod \"console-operator-58897d9998-xvd7z\" (UID: \"56b3121f-1c64-40d2-9cfd-b2c4a55451cc\") " pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.228669 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.247580 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.252546 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.253972 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" event={"ID":"018cb267-3656-4202-ba7e-323811533a19","Type":"ContainerStarted","Data":"ff2c570d5bf652033cdf52ea20f44fe2d14fb75d8120fee4a903ceeb3721f91e"} Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.258368 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" event={"ID":"0095d26a-f43b-4d74-a34e-f5a3bf289dfd","Type":"ContainerStarted","Data":"c2a7777083ef1b62bca83e11c3fad22d13876c6fc2b33afb14ed5cc47ee105ad"} Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.260723 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" event={"ID":"9fabab36-ee1e-4d92-b99c-b07ae87891be","Type":"ContainerStarted","Data":"2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca"} Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.260775 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" event={"ID":"9fabab36-ee1e-4d92-b99c-b07ae87891be","Type":"ContainerStarted","Data":"0a9b4bcbc351a7a5c77d7f232fb662559f925515ba44f307563e9a0b6ec04324"} Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.261776 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.265501 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.266317 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" event={"ID":"a3981843-fd55-41a8-8377-a741fcd231fb","Type":"ContainerStarted","Data":"21544cd61b5b34562039e3b1fee260249bcd0faf29482405de90e86ac9d286f2"} Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.267437 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" event={"ID":"3076dcdb-8618-4518-bf4a-bcf6b023cf0c","Type":"ContainerStarted","Data":"f8ee7aaa14773d5fe56fd64c16dfda4abb1c02879d783954fe1781161a285359"} Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.268102 4799 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xkv5k container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.268141 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" podUID="9fabab36-ee1e-4d92-b99c-b07ae87891be" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.281039 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tml2q\" (UniqueName: \"kubernetes.io/projected/c162443e-089d-49e2-968d-c712bf5f5aba-kube-api-access-tml2q\") pod \"olm-operator-6b444d44fb-8mrvt\" (UID: \"c162443e-089d-49e2-968d-c712bf5f5aba\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.287397 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.301558 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-clzwz"] Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.313640 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.320104 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.325972 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.332951 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.334903 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7f9sl"] Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.347762 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.366473 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.406135 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.413726 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.425717 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.428748 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:21:59 crc kubenswrapper[4799]: W0930 14:21:59.442835 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a56ad12_0b83_4023_b2d1_b3b6f9e5f1ea.slice/crio-646c6fe8a6d2c6574d64f7f085adddc9e5ada159489857ca017b0ae7434f32fa WatchSource:0}: Error finding container 646c6fe8a6d2c6574d64f7f085adddc9e5ada159489857ca017b0ae7434f32fa: Status 404 returned error can't find the container with id 646c6fe8a6d2c6574d64f7f085adddc9e5ada159489857ca017b0ae7434f32fa Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.448193 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.453456 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.465934 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.486959 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.507629 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.520279 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5"] Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.529258 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.543958 4799 request.go:700] Waited for 1.80669762s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.546400 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.573568 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.586518 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.607093 4799 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.626975 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.669984 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.675872 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/383dabba-8b1c-4e84-945a-6d6b3749802c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-df8nt\" (UID: \"383dabba-8b1c-4e84-945a-6d6b3749802c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.686754 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.706340 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.715624 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j"] Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780195 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-trusted-ca\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780251 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-service-ca\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-trusted-ca-bundle\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780342 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8485bb78-9aed-4175-8a10-498394ad1524-proxy-tls\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-certificates\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780380 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-serving-cert\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780426 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-oauth-serving-cert\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780462 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7c75754-cd2c-436f-8eea-08e68904923b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780480 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h4c2\" (UniqueName: \"kubernetes.io/projected/e7c75754-cd2c-436f-8eea-08e68904923b-kube-api-access-9h4c2\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780506 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n8sx\" (UniqueName: \"kubernetes.io/projected/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-kube-api-access-5n8sx\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780529 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-tls\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780542 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-ca\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780575 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/04c8f827-2fda-4565-9186-b78529a4b427-metrics-tls\") pod \"dns-operator-744455d44c-dzk4q\" (UID: \"04c8f827-2fda-4565-9186-b78529a4b427\") " pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780637 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780674 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-config\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780707 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-config\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780723 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls4sr\" (UniqueName: \"kubernetes.io/projected/04c8f827-2fda-4565-9186-b78529a4b427-kube-api-access-ls4sr\") pod \"dns-operator-744455d44c-dzk4q\" (UID: \"04c8f827-2fda-4565-9186-b78529a4b427\") " pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780739 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-serving-cert\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780772 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-oauth-config\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780796 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7675ad08-1655-418f-b0c1-baac6a2d35b6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780810 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm2jz\" (UniqueName: \"kubernetes.io/projected/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-kube-api-access-rm2jz\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780825 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7c75754-cd2c-436f-8eea-08e68904923b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780841 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7232c443-19e4-4ccd-af54-725f6d10d808-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780863 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-serving-cert\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780910 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-config\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780940 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8485bb78-9aed-4175-8a10-498394ad1524-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780955 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-service-ca\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780978 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7c75754-cd2c-436f-8eea-08e68904923b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.780993 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7232c443-19e4-4ccd-af54-725f6d10d808-config\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781082 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-service-ca-bundle\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781098 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7232c443-19e4-4ccd-af54-725f6d10d808-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781114 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k779\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-kube-api-access-9k779\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781149 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-client\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781172 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7675ad08-1655-418f-b0c1-baac6a2d35b6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781188 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l8v6\" (UniqueName: \"kubernetes.io/projected/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-kube-api-access-2l8v6\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxgwp\" (UniqueName: \"kubernetes.io/projected/8485bb78-9aed-4175-8a10-498394ad1524-kube-api-access-rxgwp\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781218 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.781232 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-bound-sa-token\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: E0930 14:21:59.786370 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.286353027 +0000 UTC m=+142.369953534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.818173 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz"] Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.862844 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882342 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882559 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls4sr\" (UniqueName: \"kubernetes.io/projected/04c8f827-2fda-4565-9186-b78529a4b427-kube-api-access-ls4sr\") pod \"dns-operator-744455d44c-dzk4q\" (UID: \"04c8f827-2fda-4565-9186-b78529a4b427\") " pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882592 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-serving-cert\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882614 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-oauth-config\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882641 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a317838-165c-44ec-a74e-8fa0ebaa70d6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:21:59 crc kubenswrapper[4799]: E0930 14:21:59.882866 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.382799722 +0000 UTC m=+142.466400149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882909 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7675ad08-1655-418f-b0c1-baac6a2d35b6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882933 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7c75754-cd2c-436f-8eea-08e68904923b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882963 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7232c443-19e4-4ccd-af54-725f6d10d808-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.882989 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6pmd\" (UniqueName: \"kubernetes.io/projected/ae4eb023-1b23-4106-9203-9326737be923-kube-api-access-j6pmd\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883014 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbr2l\" (UniqueName: \"kubernetes.io/projected/2e235578-9f56-4b94-aa1b-ad79bf0f8957-kube-api-access-zbr2l\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883074 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8485bb78-9aed-4175-8a10-498394ad1524-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883100 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-service-ca\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883143 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c44311c-64e8-4587-95c6-2bdda1c93ad1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5jqc\" (UID: \"7c44311c-64e8-4587-95c6-2bdda1c93ad1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883168 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-service-ca-bundle\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883219 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k779\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-kube-api-access-9k779\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883241 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-stats-auth\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883263 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fd8f596-14c4-4269-b755-b28a70c69647-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8wmvv\" (UID: \"6fd8f596-14c4-4269-b755-b28a70c69647\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883278 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjmzh\" (UniqueName: \"kubernetes.io/projected/367e05f1-fb1d-435b-978a-83486f04ada1-kube-api-access-zjmzh\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883295 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-socket-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.883352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l8v6\" (UniqueName: \"kubernetes.io/projected/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-kube-api-access-2l8v6\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.884218 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-service-ca\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.885491 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-service-ca-bundle\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.886707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxgwp\" (UniqueName: \"kubernetes.io/projected/8485bb78-9aed-4175-8a10-498394ad1524-kube-api-access-rxgwp\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.886808 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.887861 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjrff\" (UniqueName: \"kubernetes.io/projected/2fa0f587-c724-4534-888a-33bc5f77f56a-kube-api-access-tjrff\") pod \"ingress-canary-jznkv\" (UID: \"2fa0f587-c724-4534-888a-33bc5f77f56a\") " pod="openshift-ingress-canary/ingress-canary-jznkv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.887893 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4n5p\" (UniqueName: \"kubernetes.io/projected/6e587e9f-8d37-46ef-ab7e-23e5f772078b-kube-api-access-m4n5p\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888019 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/367e05f1-fb1d-435b-978a-83486f04ada1-signing-key\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888044 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34913ce8-3c56-4325-a430-bd2faa213576-webhook-cert\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888061 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/34913ce8-3c56-4325-a430-bd2faa213576-tmpfs\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-metrics-certs\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888216 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fa0f587-c724-4534-888a-33bc5f77f56a-cert\") pod \"ingress-canary-jznkv\" (UID: \"2fa0f587-c724-4534-888a-33bc5f77f56a\") " pod="openshift-ingress-canary/ingress-canary-jznkv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888357 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-certificates\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888559 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-oauth-serving-cert\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888598 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h4c2\" (UniqueName: \"kubernetes.io/projected/e7c75754-cd2c-436f-8eea-08e68904923b-kube-api-access-9h4c2\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n8sx\" (UniqueName: \"kubernetes.io/projected/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-kube-api-access-5n8sx\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.888635 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-tls\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-ca\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889105 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a317838-165c-44ec-a74e-8fa0ebaa70d6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889120 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889137 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpbqb\" (UniqueName: \"kubernetes.io/projected/d931fb63-da78-4311-ba16-e5b931a12dad-kube-api-access-bpbqb\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889166 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae4eb023-1b23-4106-9203-9326737be923-config-volume\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889187 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889204 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnrrg\" (UniqueName: \"kubernetes.io/projected/292776df-8fba-4f59-bcfa-e48d6bb5626d-kube-api-access-wnrrg\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889220 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d931fb63-da78-4311-ba16-e5b931a12dad-serving-cert\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889243 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/292776df-8fba-4f59-bcfa-e48d6bb5626d-node-bootstrap-token\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889258 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889484 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.887270 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-oauth-config\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.887574 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7675ad08-1655-418f-b0c1-baac6a2d35b6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889836 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a317838-165c-44ec-a74e-8fa0ebaa70d6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889923 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-config\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889947 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvczf\" (UniqueName: \"kubernetes.io/projected/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-kube-api-access-tvczf\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889964 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fbt2\" (UniqueName: \"kubernetes.io/projected/34913ce8-3c56-4325-a430-bd2faa213576-kube-api-access-7fbt2\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.889982 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm2jz\" (UniqueName: \"kubernetes.io/projected/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-kube-api-access-rm2jz\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890001 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-serving-cert\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890018 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-proxy-tls\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890033 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-plugins-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890061 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-config\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890076 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-images\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890091 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h775b\" (UniqueName: \"kubernetes.io/projected/7c44311c-64e8-4587-95c6-2bdda1c93ad1-kube-api-access-h775b\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5jqc\" (UID: \"7c44311c-64e8-4587-95c6-2bdda1c93ad1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7c75754-cd2c-436f-8eea-08e68904923b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7232c443-19e4-4ccd-af54-725f6d10d808-config\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890141 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-default-certificate\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890167 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5rcq\" (UniqueName: \"kubernetes.io/projected/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-kube-api-access-k5rcq\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890193 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7232c443-19e4-4ccd-af54-725f6d10d808-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890214 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-csi-data-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890246 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f147ef4-6fba-40b2-bdda-f19618d512eb-config-volume\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890261 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qxxn\" (UniqueName: \"kubernetes.io/projected/7f147ef4-6fba-40b2-bdda-f19618d512eb-kube-api-access-9qxxn\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890286 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-client\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7675ad08-1655-418f-b0c1-baac6a2d35b6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890316 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-registration-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890330 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z9r4\" (UniqueName: \"kubernetes.io/projected/99ec40a6-39fc-44cb-bcc6-39bd7e7019b8-kube-api-access-7z9r4\") pod \"migrator-59844c95c7-p55n7\" (UID: \"99ec40a6-39fc-44cb-bcc6-39bd7e7019b8\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890353 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3449f7db-d420-4de1-9193-65dc16c8cd45-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-77w76\" (UID: \"3449f7db-d420-4de1-9193-65dc16c8cd45\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtxs\" (UniqueName: \"kubernetes.io/projected/3449f7db-d420-4de1-9193-65dc16c8cd45-kube-api-access-qgtxs\") pod \"package-server-manager-789f6589d5-77w76\" (UID: \"3449f7db-d420-4de1-9193-65dc16c8cd45\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890390 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890403 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/292776df-8fba-4f59-bcfa-e48d6bb5626d-certs\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890416 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f147ef4-6fba-40b2-bdda-f19618d512eb-secret-volume\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890433 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-bound-sa-token\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890447 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2e235578-9f56-4b94-aa1b-ad79bf0f8957-srv-cert\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890460 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-mountpoint-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890476 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e587e9f-8d37-46ef-ab7e-23e5f772078b-service-ca-bundle\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890494 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-trusted-ca\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890508 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8xnr\" (UniqueName: \"kubernetes.io/projected/6fd8f596-14c4-4269-b755-b28a70c69647-kube-api-access-x8xnr\") pod \"multus-admission-controller-857f4d67dd-8wmvv\" (UID: \"6fd8f596-14c4-4269-b755-b28a70c69647\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890538 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-certificates\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-service-ca\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890569 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/367e05f1-fb1d-435b-978a-83486f04ada1-signing-cabundle\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890596 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h975p\" (UniqueName: \"kubernetes.io/projected/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-kube-api-access-h975p\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-trusted-ca-bundle\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8485bb78-9aed-4175-8a10-498394ad1524-proxy-tls\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890707 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890726 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae4eb023-1b23-4106-9203-9326737be923-metrics-tls\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.890748 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-serving-cert\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.892859 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7c75754-cd2c-436f-8eea-08e68904923b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.892977 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34913ce8-3c56-4325-a430-bd2faa213576-apiservice-cert\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.893047 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/04c8f827-2fda-4565-9186-b78529a4b427-metrics-tls\") pod \"dns-operator-744455d44c-dzk4q\" (UID: \"04c8f827-2fda-4565-9186-b78529a4b427\") " pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.893100 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d931fb63-da78-4311-ba16-e5b931a12dad-config\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.893128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2e235578-9f56-4b94-aa1b-ad79bf0f8957-profile-collector-cert\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.893173 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s2cw\" (UniqueName: \"kubernetes.io/projected/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-kube-api-access-8s2cw\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.893215 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-config\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.894546 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-config\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: E0930 14:21:59.895162 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.395147444 +0000 UTC m=+142.478747871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.897669 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-config\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.898685 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-serving-cert\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.900822 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-ca\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.901739 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7675ad08-1655-418f-b0c1-baac6a2d35b6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.902234 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-trusted-ca-bundle\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.902981 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-config\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.903998 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-client\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.904282 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-trusted-ca\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.904578 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-oauth-serving-cert\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.905150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-etcd-service-ca\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.910293 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7232c443-19e4-4ccd-af54-725f6d10d808-config\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.912003 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7c75754-cd2c-436f-8eea-08e68904923b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.912225 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7232c443-19e4-4ccd-af54-725f6d10d808-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.912551 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-serving-cert\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.912932 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-serving-cert\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.916263 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7c75754-cd2c-436f-8eea-08e68904923b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.919133 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8485bb78-9aed-4175-8a10-498394ad1524-proxy-tls\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.920861 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-tls\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.926519 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/04c8f827-2fda-4565-9186-b78529a4b427-metrics-tls\") pod \"dns-operator-744455d44c-dzk4q\" (UID: \"04c8f827-2fda-4565-9186-b78529a4b427\") " pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.927484 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls4sr\" (UniqueName: \"kubernetes.io/projected/04c8f827-2fda-4565-9186-b78529a4b427-kube-api-access-ls4sr\") pod \"dns-operator-744455d44c-dzk4q\" (UID: \"04c8f827-2fda-4565-9186-b78529a4b427\") " pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.940743 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7c75754-cd2c-436f-8eea-08e68904923b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.960219 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7232c443-19e4-4ccd-af54-725f6d10d808-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fsfrv\" (UID: \"7232c443-19e4-4ccd-af54-725f6d10d808\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.981733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l8v6\" (UniqueName: \"kubernetes.io/projected/3e46f955-630b-4fc7-ab88-0d8fcaddc68c-kube-api-access-2l8v6\") pod \"etcd-operator-b45778765-2tlvs\" (UID: \"3e46f955-630b-4fc7-ab88-0d8fcaddc68c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.994563 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:21:59 crc kubenswrapper[4799]: E0930 14:21:59.994908 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.494890554 +0000 UTC m=+142.578490981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.995094 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.995123 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae4eb023-1b23-4106-9203-9326737be923-metrics-tls\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.996047 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.996834 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34913ce8-3c56-4325-a430-bd2faa213576-apiservice-cert\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.996913 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d931fb63-da78-4311-ba16-e5b931a12dad-config\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.996938 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2e235578-9f56-4b94-aa1b-ad79bf0f8957-profile-collector-cert\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.996960 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s2cw\" (UniqueName: \"kubernetes.io/projected/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-kube-api-access-8s2cw\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.996988 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a317838-165c-44ec-a74e-8fa0ebaa70d6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6pmd\" (UniqueName: \"kubernetes.io/projected/ae4eb023-1b23-4106-9203-9326737be923-kube-api-access-j6pmd\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997029 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbr2l\" (UniqueName: \"kubernetes.io/projected/2e235578-9f56-4b94-aa1b-ad79bf0f8957-kube-api-access-zbr2l\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997067 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c44311c-64e8-4587-95c6-2bdda1c93ad1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5jqc\" (UID: \"7c44311c-64e8-4587-95c6-2bdda1c93ad1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997087 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-stats-auth\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997126 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fd8f596-14c4-4269-b755-b28a70c69647-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8wmvv\" (UID: \"6fd8f596-14c4-4269-b755-b28a70c69647\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997142 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjmzh\" (UniqueName: \"kubernetes.io/projected/367e05f1-fb1d-435b-978a-83486f04ada1-kube-api-access-zjmzh\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997158 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-socket-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997187 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjrff\" (UniqueName: \"kubernetes.io/projected/2fa0f587-c724-4534-888a-33bc5f77f56a-kube-api-access-tjrff\") pod \"ingress-canary-jznkv\" (UID: \"2fa0f587-c724-4534-888a-33bc5f77f56a\") " pod="openshift-ingress-canary/ingress-canary-jznkv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997207 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/367e05f1-fb1d-435b-978a-83486f04ada1-signing-key\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997221 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4n5p\" (UniqueName: \"kubernetes.io/projected/6e587e9f-8d37-46ef-ab7e-23e5f772078b-kube-api-access-m4n5p\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997236 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34913ce8-3c56-4325-a430-bd2faa213576-webhook-cert\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997262 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/34913ce8-3c56-4325-a430-bd2faa213576-tmpfs\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997323 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-metrics-certs\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997341 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fa0f587-c724-4534-888a-33bc5f77f56a-cert\") pod \"ingress-canary-jznkv\" (UID: \"2fa0f587-c724-4534-888a-33bc5f77f56a\") " pod="openshift-ingress-canary/ingress-canary-jznkv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a317838-165c-44ec-a74e-8fa0ebaa70d6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997431 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpbqb\" (UniqueName: \"kubernetes.io/projected/d931fb63-da78-4311-ba16-e5b931a12dad-kube-api-access-bpbqb\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae4eb023-1b23-4106-9203-9326737be923-config-volume\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997465 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnrrg\" (UniqueName: \"kubernetes.io/projected/292776df-8fba-4f59-bcfa-e48d6bb5626d-kube-api-access-wnrrg\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d931fb63-da78-4311-ba16-e5b931a12dad-serving-cert\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997502 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997521 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997540 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a317838-165c-44ec-a74e-8fa0ebaa70d6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997562 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/292776df-8fba-4f59-bcfa-e48d6bb5626d-node-bootstrap-token\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997592 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvczf\" (UniqueName: \"kubernetes.io/projected/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-kube-api-access-tvczf\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997614 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fbt2\" (UniqueName: \"kubernetes.io/projected/34913ce8-3c56-4325-a430-bd2faa213576-kube-api-access-7fbt2\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997642 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-proxy-tls\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997703 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-plugins-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997720 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-images\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997746 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h775b\" (UniqueName: \"kubernetes.io/projected/7c44311c-64e8-4587-95c6-2bdda1c93ad1-kube-api-access-h775b\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5jqc\" (UID: \"7c44311c-64e8-4587-95c6-2bdda1c93ad1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997772 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-default-certificate\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997795 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5rcq\" (UniqueName: \"kubernetes.io/projected/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-kube-api-access-k5rcq\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997817 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-csi-data-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997835 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f147ef4-6fba-40b2-bdda-f19618d512eb-config-volume\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997860 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qxxn\" (UniqueName: \"kubernetes.io/projected/7f147ef4-6fba-40b2-bdda-f19618d512eb-kube-api-access-9qxxn\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997885 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-registration-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.997907 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z9r4\" (UniqueName: \"kubernetes.io/projected/99ec40a6-39fc-44cb-bcc6-39bd7e7019b8-kube-api-access-7z9r4\") pod \"migrator-59844c95c7-p55n7\" (UID: \"99ec40a6-39fc-44cb-bcc6-39bd7e7019b8\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998014 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3449f7db-d420-4de1-9193-65dc16c8cd45-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-77w76\" (UID: \"3449f7db-d420-4de1-9193-65dc16c8cd45\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998042 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgtxs\" (UniqueName: \"kubernetes.io/projected/3449f7db-d420-4de1-9193-65dc16c8cd45-kube-api-access-qgtxs\") pod \"package-server-manager-789f6589d5-77w76\" (UID: \"3449f7db-d420-4de1-9193-65dc16c8cd45\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998066 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/292776df-8fba-4f59-bcfa-e48d6bb5626d-certs\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998087 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998113 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f147ef4-6fba-40b2-bdda-f19618d512eb-secret-volume\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998166 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-mountpoint-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998185 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2e235578-9f56-4b94-aa1b-ad79bf0f8957-srv-cert\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998202 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e587e9f-8d37-46ef-ab7e-23e5f772078b-service-ca-bundle\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998229 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8xnr\" (UniqueName: \"kubernetes.io/projected/6fd8f596-14c4-4269-b755-b28a70c69647-kube-api-access-x8xnr\") pod \"multus-admission-controller-857f4d67dd-8wmvv\" (UID: \"6fd8f596-14c4-4269-b755-b28a70c69647\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998260 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/367e05f1-fb1d-435b-978a-83486f04ada1-signing-cabundle\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h975p\" (UniqueName: \"kubernetes.io/projected/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-kube-api-access-h975p\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998803 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/34913ce8-3c56-4325-a430-bd2faa213576-tmpfs\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.998955 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae4eb023-1b23-4106-9203-9326737be923-metrics-tls\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.999268 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k779\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-kube-api-access-9k779\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.999506 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-socket-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:21:59 crc kubenswrapper[4799]: I0930 14:21:59.999559 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.000014 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34913ce8-3c56-4325-a430-bd2faa213576-apiservice-cert\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.001803 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-metrics-certs\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.001823 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34913ce8-3c56-4325-a430-bd2faa213576-webhook-cert\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.002586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-default-certificate\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.002786 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-csi-data-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.003427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f147ef4-6fba-40b2-bdda-f19618d512eb-config-volume\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.003555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-registration-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.004531 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6fd8f596-14c4-4269-b755-b28a70c69647-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8wmvv\" (UID: \"6fd8f596-14c4-4269-b755-b28a70c69647\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.004638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/367e05f1-fb1d-435b-978a-83486f04ada1-signing-key\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.006371 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fa0f587-c724-4534-888a-33bc5f77f56a-cert\") pod \"ingress-canary-jznkv\" (UID: \"2fa0f587-c724-4534-888a-33bc5f77f56a\") " pod="openshift-ingress-canary/ingress-canary-jznkv" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.006995 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3449f7db-d420-4de1-9193-65dc16c8cd45-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-77w76\" (UID: \"3449f7db-d420-4de1-9193-65dc16c8cd45\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.010127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a317838-165c-44ec-a74e-8fa0ebaa70d6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.010804 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.016535 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.516519041 +0000 UTC m=+142.600119468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.016593 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-plugins-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.017199 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae4eb023-1b23-4106-9203-9326737be923-config-volume\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.017737 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/367e05f1-fb1d-435b-978a-83486f04ada1-signing-cabundle\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.018126 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/292776df-8fba-4f59-bcfa-e48d6bb5626d-certs\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.018942 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d931fb63-da78-4311-ba16-e5b931a12dad-serving-cert\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.020339 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-mountpoint-dir\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.021795 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-images\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.023753 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f147ef4-6fba-40b2-bdda-f19618d512eb-secret-volume\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.024069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/292776df-8fba-4f59-bcfa-e48d6bb5626d-node-bootstrap-token\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.024095 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-proxy-tls\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.026261 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxgwp\" (UniqueName: \"kubernetes.io/projected/8485bb78-9aed-4175-8a10-498394ad1524-kube-api-access-rxgwp\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.026417 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c44311c-64e8-4587-95c6-2bdda1c93ad1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5jqc\" (UID: \"7c44311c-64e8-4587-95c6-2bdda1c93ad1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.036931 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xvd7z"] Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.044495 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h4c2\" (UniqueName: \"kubernetes.io/projected/e7c75754-cd2c-436f-8eea-08e68904923b-kube-api-access-9h4c2\") pod \"cluster-image-registry-operator-dc59b4c8b-bdngs\" (UID: \"e7c75754-cd2c-436f-8eea-08e68904923b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.049171 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz"] Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.061617 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt"] Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.063568 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n8sx\" (UniqueName: \"kubernetes.io/projected/e81f8e3f-96d2-4aa6-9a74-765d80da36b5-kube-api-access-5n8sx\") pod \"authentication-operator-69f744f599-w9q9n\" (UID: \"e81f8e3f-96d2-4aa6-9a74-765d80da36b5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.080508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm2jz\" (UniqueName: \"kubernetes.io/projected/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-kube-api-access-rm2jz\") pod \"console-f9d7485db-svk57\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.081933 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.082265 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.082688 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d931fb63-da78-4311-ba16-e5b931a12dad-config\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.083433 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8485bb78-9aed-4175-8a10-498394ad1524-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cvsnb\" (UID: \"8485bb78-9aed-4175-8a10-498394ad1524\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.083591 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a317838-165c-44ec-a74e-8fa0ebaa70d6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.084249 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6e587e9f-8d37-46ef-ab7e-23e5f772078b-stats-auth\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.084472 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e587e9f-8d37-46ef-ab7e-23e5f772078b-service-ca-bundle\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.085400 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2e235578-9f56-4b94-aa1b-ad79bf0f8957-profile-collector-cert\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:22:00 crc kubenswrapper[4799]: W0930 14:22:00.088030 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56b3121f_1c64_40d2_9cfd_b2c4a55451cc.slice/crio-180d8c5856bf15aa501303e507640dc0975b8fa23554054216e6a6c9ba311a66 WatchSource:0}: Error finding container 180d8c5856bf15aa501303e507640dc0975b8fa23554054216e6a6c9ba311a66: Status 404 returned error can't find the container with id 180d8c5856bf15aa501303e507640dc0975b8fa23554054216e6a6c9ba311a66 Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.088964 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2e235578-9f56-4b94-aa1b-ad79bf0f8957-srv-cert\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.098870 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.099373 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.599356528 +0000 UTC m=+142.682956955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.121243 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.128058 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-bound-sa-token\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.147171 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.150767 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h975p\" (UniqueName: \"kubernetes.io/projected/d119c9e7-f611-40ef-8c19-fb5a1ed9cf64-kube-api-access-h975p\") pod \"csi-hostpathplugin-mkz8h\" (UID: \"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.177870 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbr2l\" (UniqueName: \"kubernetes.io/projected/2e235578-9f56-4b94-aa1b-ad79bf0f8957-kube-api-access-zbr2l\") pod \"catalog-operator-68c6474976-fzmhc\" (UID: \"2e235578-9f56-4b94-aa1b-ad79bf0f8957\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.187987 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.192068 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.192606 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6pmd\" (UniqueName: \"kubernetes.io/projected/ae4eb023-1b23-4106-9203-9326737be923-kube-api-access-j6pmd\") pod \"dns-default-vxm25\" (UID: \"ae4eb023-1b23-4106-9203-9326737be923\") " pod="openshift-dns/dns-default-vxm25" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.203689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.204054 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.704042448 +0000 UTC m=+142.787642875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.206325 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgtxs\" (UniqueName: \"kubernetes.io/projected/3449f7db-d420-4de1-9193-65dc16c8cd45-kube-api-access-qgtxs\") pod \"package-server-manager-789f6589d5-77w76\" (UID: \"3449f7db-d420-4de1-9193-65dc16c8cd45\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.216486 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.225790 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h775b\" (UniqueName: \"kubernetes.io/projected/7c44311c-64e8-4587-95c6-2bdda1c93ad1-kube-api-access-h775b\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5jqc\" (UID: \"7c44311c-64e8-4587-95c6-2bdda1c93ad1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.225847 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.242005 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4n5p\" (UniqueName: \"kubernetes.io/projected/6e587e9f-8d37-46ef-ab7e-23e5f772078b-kube-api-access-m4n5p\") pod \"router-default-5444994796-glzch\" (UID: \"6e587e9f-8d37-46ef-ab7e-23e5f772078b\") " pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.242221 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.252079 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.264712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5rcq\" (UniqueName: \"kubernetes.io/projected/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-kube-api-access-k5rcq\") pod \"marketplace-operator-79b997595-b7pv9\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.273427 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" event={"ID":"018cb267-3656-4202-ba7e-323811533a19","Type":"ContainerStarted","Data":"022f9fddca47391a4030178c97c90a621bb7b3c4971b6ad33567f6d4f7b64c96"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.274438 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" event={"ID":"d8918fed-2347-4360-a4e1-d84658a2211f","Type":"ContainerStarted","Data":"1351b3b33d1b317f5b67a0ab8929ea7fd2e88f8a2661e8c35f12049faa0039b7"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.276045 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.276214 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" event={"ID":"c162443e-089d-49e2-968d-c712bf5f5aba","Type":"ContainerStarted","Data":"43ee47975a48efadf39fff86e82d55db0a474375e90e3af767a8435a82277c69"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.278440 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" event={"ID":"3076dcdb-8618-4518-bf4a-bcf6b023cf0c","Type":"ContainerStarted","Data":"3f0d058258bddbded570425f40bf612215e49af6ecdbbc295f29464530667834"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.279495 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" event={"ID":"56b3121f-1c64-40d2-9cfd-b2c4a55451cc","Type":"ContainerStarted","Data":"180d8c5856bf15aa501303e507640dc0975b8fa23554054216e6a6c9ba311a66"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.283008 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.283269 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qxxn\" (UniqueName: \"kubernetes.io/projected/7f147ef4-6fba-40b2-bdda-f19618d512eb-kube-api-access-9qxxn\") pod \"collect-profiles-29320695-r4v2t\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.287899 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" event={"ID":"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea","Type":"ContainerStarted","Data":"646c6fe8a6d2c6574d64f7f085adddc9e5ada159489857ca017b0ae7434f32fa"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.299835 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" event={"ID":"dedc044b-a09f-436a-8bdb-3b606269e3c5","Type":"ContainerStarted","Data":"368cfcf7c9f760af45b78a9cbb3403ba08e1b7fc88101094f6d09ca3ddd4b084"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.300570 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt"] Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.302231 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z9r4\" (UniqueName: \"kubernetes.io/projected/99ec40a6-39fc-44cb-bcc6-39bd7e7019b8-kube-api-access-7z9r4\") pod \"migrator-59844c95c7-p55n7\" (UID: \"99ec40a6-39fc-44cb-bcc6-39bd7e7019b8\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.302531 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" event={"ID":"3d01e81a-5012-4d30-863b-49396073cbd1","Type":"ContainerStarted","Data":"7bd046556f6056c70e24f37c9048ab9fbe97337df2afc0045be0d6e4769aabc2"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.304526 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.304667 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.804635252 +0000 UTC m=+142.888235679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.304859 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.305177 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.805163577 +0000 UTC m=+142.888764004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.306360 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" event={"ID":"a2212e6e-a106-4efc-bb26-27bb27f8c0b8","Type":"ContainerStarted","Data":"7a9cb359afe19c951b7dc6d6fadc3c9cc9201a0f8f7f3de4b04e6145d35e3916"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.307547 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" event={"ID":"c35fadda-30b5-438e-b0a6-945004f36333","Type":"ContainerStarted","Data":"14779a4953735ff9ef866ce4d6a55adaead399af35abcfeb91167287895d7c4d"} Sep 30 14:22:00 crc kubenswrapper[4799]: W0930 14:22:00.311251 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod383dabba_8b1c_4e84_945a_6d6b3749802c.slice/crio-1db270a4db701c9be6e57100f084085435b4e9e6396530b2d1ba616cd6c55d68 WatchSource:0}: Error finding container 1db270a4db701c9be6e57100f084085435b4e9e6396530b2d1ba616cd6c55d68: Status 404 returned error can't find the container with id 1db270a4db701c9be6e57100f084085435b4e9e6396530b2d1ba616cd6c55d68 Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.313041 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.315952 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" event={"ID":"0095d26a-f43b-4d74-a34e-f5a3bf289dfd","Type":"ContainerStarted","Data":"06caa9bba371f110e564867dfe787f4e5fb08e8a22d6653bfc045f4c9980af26"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.317369 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" event={"ID":"8209103b-69e0-41be-9ce9-c87dd3e01a19","Type":"ContainerStarted","Data":"c859467e8e38726e4c6f47827c40d004725b0e44010c54b194e56ef0c3175c54"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.319438 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjrff\" (UniqueName: \"kubernetes.io/projected/2fa0f587-c724-4534-888a-33bc5f77f56a-kube-api-access-tjrff\") pod \"ingress-canary-jznkv\" (UID: \"2fa0f587-c724-4534-888a-33bc5f77f56a\") " pod="openshift-ingress-canary/ingress-canary-jznkv" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.321620 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" event={"ID":"a3981843-fd55-41a8-8377-a741fcd231fb","Type":"ContainerStarted","Data":"0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.322613 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.325035 4799 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rmw7h container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.325091 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" podUID="a3981843-fd55-41a8-8377-a741fcd231fb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.327020 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7f9sl" event={"ID":"e2e2061e-3d24-41dd-b8ee-6b50e8313647","Type":"ContainerStarted","Data":"0fb9f012503196c07acd46c85420f5f64dbd0054be968eaac1851ba85a316966"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.327053 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7f9sl" event={"ID":"e2e2061e-3d24-41dd-b8ee-6b50e8313647","Type":"ContainerStarted","Data":"d99464ee6abe9c2195d7ff8f8cbdda27603c5ec5fba16e6e33c317494304ca3c"} Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.327726 4799 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xkv5k container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.327765 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" podUID="9fabab36-ee1e-4d92-b99c-b07ae87891be" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.345028 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.357047 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.359594 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8a317838-165c-44ec-a74e-8fa0ebaa70d6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cp7sz\" (UID: \"8a317838-165c-44ec-a74e-8fa0ebaa70d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.366905 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s2cw\" (UniqueName: \"kubernetes.io/projected/26b28ca2-e7b1-4e6e-ba95-ac537580f7c1-kube-api-access-8s2cw\") pod \"machine-config-operator-74547568cd-xgs5n\" (UID: \"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.368066 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.377499 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jznkv" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.395347 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvczf\" (UniqueName: \"kubernetes.io/projected/99584c4c-3c3d-4bae-99c0-2eccda0b27c0-kube-api-access-tvczf\") pod \"kube-storage-version-migrator-operator-b67b599dd-xhs4m\" (UID: \"99584c4c-3c3d-4bae-99c0-2eccda0b27c0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.410879 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.411357 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fbt2\" (UniqueName: \"kubernetes.io/projected/34913ce8-3c56-4325-a430-bd2faa213576-kube-api-access-7fbt2\") pod \"packageserver-d55dfcdfc-jj9bm\" (UID: \"34913ce8-3c56-4325-a430-bd2faa213576\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.411527 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.412954 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:00.912935716 +0000 UTC m=+142.996536143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.413814 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vxm25" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.426737 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8xnr\" (UniqueName: \"kubernetes.io/projected/6fd8f596-14c4-4269-b755-b28a70c69647-kube-api-access-x8xnr\") pod \"multus-admission-controller-857f4d67dd-8wmvv\" (UID: \"6fd8f596-14c4-4269-b755-b28a70c69647\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.442672 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpbqb\" (UniqueName: \"kubernetes.io/projected/d931fb63-da78-4311-ba16-e5b931a12dad-kube-api-access-bpbqb\") pod \"service-ca-operator-777779d784-z7hn5\" (UID: \"d931fb63-da78-4311-ba16-e5b931a12dad\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.471428 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnrrg\" (UniqueName: \"kubernetes.io/projected/292776df-8fba-4f59-bcfa-e48d6bb5626d-kube-api-access-wnrrg\") pod \"machine-config-server-m4jz9\" (UID: \"292776df-8fba-4f59-bcfa-e48d6bb5626d\") " pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.512745 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjmzh\" (UniqueName: \"kubernetes.io/projected/367e05f1-fb1d-435b-978a-83486f04ada1-kube-api-access-zjmzh\") pod \"service-ca-9c57cc56f-hrm8v\" (UID: \"367e05f1-fb1d-435b-978a-83486f04ada1\") " pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.513901 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.514992 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.014976661 +0000 UTC m=+143.098577088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.558813 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.570919 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.590922 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.603391 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.605128 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.618286 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2tlvs"] Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.622060 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.622512 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.122460661 +0000 UTC m=+143.206061098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.622600 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.639929 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.686957 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-m4jz9" Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.723234 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.723548 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.223536288 +0000 UTC m=+143.307136715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.824227 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.825108 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.325092239 +0000 UTC m=+143.408692666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.885672 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w9q9n"] Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.903349 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs"] Sep 30 14:22:00 crc kubenswrapper[4799]: I0930 14:22:00.926020 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:00 crc kubenswrapper[4799]: E0930 14:22:00.927967 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.427951728 +0000 UTC m=+143.511552155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.000066 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb"] Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.012931 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-svk57"] Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.033611 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.034098 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.534076599 +0000 UTC m=+143.617677026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.056003 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv"] Sep 30 14:22:01 crc kubenswrapper[4799]: W0930 14:22:01.104020 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8485bb78_9aed_4175_8a10_498394ad1524.slice/crio-0bf7674d897e720543b80d6545c38ea76d4f4335a68b3622d51a7921bf79c4dd WatchSource:0}: Error finding container 0bf7674d897e720543b80d6545c38ea76d4f4335a68b3622d51a7921bf79c4dd: Status 404 returned error can't find the container with id 0bf7674d897e720543b80d6545c38ea76d4f4335a68b3622d51a7921bf79c4dd Sep 30 14:22:01 crc kubenswrapper[4799]: W0930 14:22:01.109407 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod292776df_8fba_4f59_bcfa_e48d6bb5626d.slice/crio-dd4192f709385e812ca9953053ea2a873806621625ae95375ca0baf5bd8b31e7 WatchSource:0}: Error finding container dd4192f709385e812ca9953053ea2a873806621625ae95375ca0baf5bd8b31e7: Status 404 returned error can't find the container with id dd4192f709385e812ca9953053ea2a873806621625ae95375ca0baf5bd8b31e7 Sep 30 14:22:01 crc kubenswrapper[4799]: W0930 14:22:01.136918 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7232c443_19e4_4ccd_af54_725f6d10d808.slice/crio-ddb68940667b16d8096aaf95be8510915c5ce303e4a796002bf212302ca9af6f WatchSource:0}: Error finding container ddb68940667b16d8096aaf95be8510915c5ce303e4a796002bf212302ca9af6f: Status 404 returned error can't find the container with id ddb68940667b16d8096aaf95be8510915c5ce303e4a796002bf212302ca9af6f Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.137735 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.138121 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.638104801 +0000 UTC m=+143.721705228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.238241 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.238624 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.738605972 +0000 UTC m=+143.822206409 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.257428 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc"] Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.295426 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc"] Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.339306 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.354077 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.854056519 +0000 UTC m=+143.937656946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.413576 4799 generic.go:334] "Generic (PLEG): container finished" podID="7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea" containerID="7ae74d09bdd41d63a8e8a55bd9521636afabdd805eebfbb3b77212a1c54b1bf8" exitCode=0 Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.414342 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" event={"ID":"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea","Type":"ContainerDied","Data":"7ae74d09bdd41d63a8e8a55bd9521636afabdd805eebfbb3b77212a1c54b1bf8"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.434493 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dzk4q"] Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.444391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.445149 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:01.94512967 +0000 UTC m=+144.028730097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.456073 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" event={"ID":"c35fadda-30b5-438e-b0a6-945004f36333","Type":"ContainerStarted","Data":"80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.456405 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.469841 4799 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bfxv2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.469894 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" podUID="c35fadda-30b5-438e-b0a6-945004f36333" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.488110 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" event={"ID":"a2212e6e-a106-4efc-bb26-27bb27f8c0b8","Type":"ContainerStarted","Data":"292f159d0a444fc8ef924883025cdd77ad0433f2d34622cc349d41a0fdc347f5"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.503903 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" event={"ID":"7232c443-19e4-4ccd-af54-725f6d10d808","Type":"ContainerStarted","Data":"ddb68940667b16d8096aaf95be8510915c5ce303e4a796002bf212302ca9af6f"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.508849 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" event={"ID":"383dabba-8b1c-4e84-945a-6d6b3749802c","Type":"ContainerStarted","Data":"1db270a4db701c9be6e57100f084085435b4e9e6396530b2d1ba616cd6c55d68"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.525565 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-svk57" event={"ID":"03c9ee3f-722e-4db8-a8f5-dc359486e4f4","Type":"ContainerStarted","Data":"ebeb1f8e458db03a90e455eefe8c04d58595ba5b8d4243c8b201320a2187aba8"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.527055 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-glzch" event={"ID":"6e587e9f-8d37-46ef-ab7e-23e5f772078b","Type":"ContainerStarted","Data":"6dd35a43b85232f545f83f607b191fde55781a4f151d61e2dd714332c1cf388f"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.533213 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" event={"ID":"e81f8e3f-96d2-4aa6-9a74-765d80da36b5","Type":"ContainerStarted","Data":"830ff32be96e9bf3344dd5b366b94ccd71fcbc8749cd556864ab2cce4a5abb1e"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.535178 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" event={"ID":"2e235578-9f56-4b94-aa1b-ad79bf0f8957","Type":"ContainerStarted","Data":"18b972d1f5e12875a8e3007d530b51b9610fd075d0f43e21c1b77739e26d7bf9"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.536846 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" event={"ID":"3e46f955-630b-4fc7-ab88-0d8fcaddc68c","Type":"ContainerStarted","Data":"46dacc19fd849028694247b95d764da7dfcd36a12fc3a096d6636fe1d4802aa4"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.547974 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.548378 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" event={"ID":"e7c75754-cd2c-436f-8eea-08e68904923b","Type":"ContainerStarted","Data":"e7a3d0f1835920bbcf4309d4957e7b9f45a15c8151560b1354bf1b7ad558455b"} Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.550370 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.050348616 +0000 UTC m=+144.133949063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.563807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" event={"ID":"c162443e-089d-49e2-968d-c712bf5f5aba","Type":"ContainerStarted","Data":"1c67e2bd14c2cb5c626dc252d3197d590b94d169207858c351cd6775e67e076b"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.566166 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.569039 4799 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-8mrvt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.569081 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" podUID="c162443e-089d-49e2-968d-c712bf5f5aba" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.582525 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-m4jz9" event={"ID":"292776df-8fba-4f59-bcfa-e48d6bb5626d","Type":"ContainerStarted","Data":"dd4192f709385e812ca9953053ea2a873806621625ae95375ca0baf5bd8b31e7"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.602886 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" event={"ID":"56b3121f-1c64-40d2-9cfd-b2c4a55451cc","Type":"ContainerStarted","Data":"f23ccee17466ab6db37229fff7213321a0fd659327e9a77646abfb4a8c0c7ebd"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.603713 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.610496 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" event={"ID":"8485bb78-9aed-4175-8a10-498394ad1524","Type":"ContainerStarted","Data":"0bf7674d897e720543b80d6545c38ea76d4f4335a68b3622d51a7921bf79c4dd"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.615699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" event={"ID":"3d01e81a-5012-4d30-863b-49396073cbd1","Type":"ContainerStarted","Data":"d7c9eea3c4b940df05f16f57b4fe0bb71dc30391242c4c9c0623ceec6165f197"} Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.616359 4799 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xkv5k container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.616390 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" podUID="9fabab36-ee1e-4d92-b99c-b07ae87891be" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.616855 4799 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rmw7h container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.616874 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" podUID="a3981843-fd55-41a8-8377-a741fcd231fb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.643019 4799 patch_prober.go:28] interesting pod/console-operator-58897d9998-xvd7z container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/readyz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.645630 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" podUID="56b3121f-1c64-40d2-9cfd-b2c4a55451cc" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/readyz\": dial tcp 10.217.0.5:8443: connect: connection refused" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.648733 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.649155 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.149130807 +0000 UTC m=+144.232731234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.649317 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.650402 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.150388613 +0000 UTC m=+144.233989040 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.722338 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" podStartSLOduration=122.722319258 podStartE2EDuration="2m2.722319258s" podCreationTimestamp="2025-09-30 14:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:01.720962859 +0000 UTC m=+143.804563306" watchObservedRunningTime="2025-09-30 14:22:01.722319258 +0000 UTC m=+143.805919685" Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.750942 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.751166 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.251111261 +0000 UTC m=+144.334711698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.751599 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.752499 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.25248746 +0000 UTC m=+144.336087887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.867574 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.868587 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.368569256 +0000 UTC m=+144.452169683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:01 crc kubenswrapper[4799]: I0930 14:22:01.972028 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:01 crc kubenswrapper[4799]: E0930 14:22:01.972408 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.472394502 +0000 UTC m=+144.555994929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.003197 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" podStartSLOduration=124.003178821 podStartE2EDuration="2m4.003178821s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.002620295 +0000 UTC m=+144.086220742" watchObservedRunningTime="2025-09-30 14:22:02.003178821 +0000 UTC m=+144.086779248" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.029353 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.077989 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.078109 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.578079181 +0000 UTC m=+144.661679608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.078253 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.078560 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.578547224 +0000 UTC m=+144.662147651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.083689 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xqj7j" podStartSLOduration=124.083640929 podStartE2EDuration="2m4.083640929s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.079172892 +0000 UTC m=+144.162773319" watchObservedRunningTime="2025-09-30 14:22:02.083640929 +0000 UTC m=+144.167241356" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.131303 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" podStartSLOduration=124.13128186 podStartE2EDuration="2m4.13128186s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.126295408 +0000 UTC m=+144.209895835" watchObservedRunningTime="2025-09-30 14:22:02.13128186 +0000 UTC m=+144.214882287" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.187559 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.188095 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.688071843 +0000 UTC m=+144.771672270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.255842 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-7f9sl" podStartSLOduration=124.255821468 podStartE2EDuration="2m4.255821468s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.255365715 +0000 UTC m=+144.338966152" watchObservedRunningTime="2025-09-30 14:22:02.255821468 +0000 UTC m=+144.339421895" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.294007 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.294316 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.794303827 +0000 UTC m=+144.877904254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.301808 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" podStartSLOduration=124.301786841 podStartE2EDuration="2m4.301786841s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.294841682 +0000 UTC m=+144.378442109" watchObservedRunningTime="2025-09-30 14:22:02.301786841 +0000 UTC m=+144.385387268" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.322583 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.329536 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.347350 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p67w5" podStartSLOduration=124.347332152 podStartE2EDuration="2m4.347332152s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.342315289 +0000 UTC m=+144.425915706" watchObservedRunningTime="2025-09-30 14:22:02.347332152 +0000 UTC m=+144.430932579" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.347495 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-b7pv9"] Sep 30 14:22:02 crc kubenswrapper[4799]: W0930 14:22:02.358276 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99584c4c_3c3d_4bae_99c0_2eccda0b27c0.slice/crio-3be99f315bf0a21fd9b26f1867d80ea3565684606de8290abc0c43a698e7a7e6 WatchSource:0}: Error finding container 3be99f315bf0a21fd9b26f1867d80ea3565684606de8290abc0c43a698e7a7e6: Status 404 returned error can't find the container with id 3be99f315bf0a21fd9b26f1867d80ea3565684606de8290abc0c43a698e7a7e6 Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.369254 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.395965 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.396595 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:02.896579819 +0000 UTC m=+144.980180246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.398747 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.414263 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8wmvv"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.414899 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" podStartSLOduration=124.414861331 podStartE2EDuration="2m4.414861331s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.399926044 +0000 UTC m=+144.483526471" watchObservedRunningTime="2025-09-30 14:22:02.414861331 +0000 UTC m=+144.498461758" Sep 30 14:22:02 crc kubenswrapper[4799]: W0930 14:22:02.440787 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f147ef4_6fba_40b2_bdda_f19618d512eb.slice/crio-080b1461b14fb7b7e7794967baafbf62538826858cabdadf8a6a8811eb4272a3 WatchSource:0}: Error finding container 080b1461b14fb7b7e7794967baafbf62538826858cabdadf8a6a8811eb4272a3: Status 404 returned error can't find the container with id 080b1461b14fb7b7e7794967baafbf62538826858cabdadf8a6a8811eb4272a3 Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.448333 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-g4fl4" podStartSLOduration=124.448309506 podStartE2EDuration="2m4.448309506s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.437966391 +0000 UTC m=+144.521566838" watchObservedRunningTime="2025-09-30 14:22:02.448309506 +0000 UTC m=+144.531909933" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.492495 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.502441 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.502772 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.002760432 +0000 UTC m=+145.086360859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: W0930 14:22:02.516375 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fd8f596_14c4_4269_b755_b28a70c69647.slice/crio-910ccc469a0effd41b6962e21eb89094ca122a854fed7120df87389346e3b8ac WatchSource:0}: Error finding container 910ccc469a0effd41b6962e21eb89094ca122a854fed7120df87389346e3b8ac: Status 404 returned error can't find the container with id 910ccc469a0effd41b6962e21eb89094ca122a854fed7120df87389346e3b8ac Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.524466 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.525252 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vxm25"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.527342 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-hrm8v"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.532918 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jznkv"] Sep 30 14:22:02 crc kubenswrapper[4799]: W0930 14:22:02.575208 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae4eb023_1b23_4106_9203_9326737be923.slice/crio-799c0f42036d8d085b9b9c029c7eaab85fe60f94d7c86e53eb1c4952d732f1e9 WatchSource:0}: Error finding container 799c0f42036d8d085b9b9c029c7eaab85fe60f94d7c86e53eb1c4952d732f1e9: Status 404 returned error can't find the container with id 799c0f42036d8d085b9b9c029c7eaab85fe60f94d7c86e53eb1c4952d732f1e9 Sep 30 14:22:02 crc kubenswrapper[4799]: W0930 14:22:02.595620 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod367e05f1_fb1d_435b_978a_83486f04ada1.slice/crio-6984e092a1ea7d956846984957989cb61ae1547a2d0bbdf30b7f6896d33094e5 WatchSource:0}: Error finding container 6984e092a1ea7d956846984957989cb61ae1547a2d0bbdf30b7f6896d33094e5: Status 404 returned error can't find the container with id 6984e092a1ea7d956846984957989cb61ae1547a2d0bbdf30b7f6896d33094e5 Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.604828 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.605778 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.105747804 +0000 UTC m=+145.189348231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.606181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.606713 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.106642239 +0000 UTC m=+145.190242666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.631856 4799 generic.go:334] "Generic (PLEG): container finished" podID="d8918fed-2347-4360-a4e1-d84658a2211f" containerID="7cc2ee6f54f8feb69806a0840e099873e67a39c9f20aea3cb9869067c49cfdfe" exitCode=0 Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.631943 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" event={"ID":"d8918fed-2347-4360-a4e1-d84658a2211f","Type":"ContainerDied","Data":"7cc2ee6f54f8feb69806a0840e099873e67a39c9f20aea3cb9869067c49cfdfe"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.646701 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mkz8h"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.649265 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" event={"ID":"7232c443-19e4-4ccd-af54-725f6d10d808","Type":"ContainerStarted","Data":"77a6bc55612636444b5c5997e549bafba635ab16a0f3efe5f9005d8197669f15"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.668427 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" event={"ID":"383dabba-8b1c-4e84-945a-6d6b3749802c","Type":"ContainerStarted","Data":"550314de61563d618a9be5305cfd433d7b1a110035bde033db1ead14be716886"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.668482 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" event={"ID":"383dabba-8b1c-4e84-945a-6d6b3749802c","Type":"ContainerStarted","Data":"616f648b4cc2547e16fedfe020685820722a201b0b2d5b7d65c7b99853801e7d"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.670079 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7"] Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.681052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" event={"ID":"04c8f827-2fda-4565-9186-b78529a4b427","Type":"ContainerStarted","Data":"5baae53bc752e5e64d1f7767bd1d6a86845dc2269d28dc5f221467a6bb7b58af"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.686796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" event={"ID":"3e46f955-630b-4fc7-ab88-0d8fcaddc68c","Type":"ContainerStarted","Data":"3e2680b047f95e46d37a9c13e07b1855efb116800ccce66f509903adf6df648a"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.690309 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" event={"ID":"6fd8f596-14c4-4269-b755-b28a70c69647","Type":"ContainerStarted","Data":"910ccc469a0effd41b6962e21eb89094ca122a854fed7120df87389346e3b8ac"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.700840 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" event={"ID":"2e235578-9f56-4b94-aa1b-ad79bf0f8957","Type":"ContainerStarted","Data":"5f940942eb7d32fd0c7e927b0f2e65d384edeea7a75e7deab88189bbf1a51b0d"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.701518 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.703459 4799 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fzmhc container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.703513 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" podUID="2e235578-9f56-4b94-aa1b-ad79bf0f8957" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.706780 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fsfrv" podStartSLOduration=124.706764919 podStartE2EDuration="2m4.706764919s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.703916718 +0000 UTC m=+144.787517155" watchObservedRunningTime="2025-09-30 14:22:02.706764919 +0000 UTC m=+144.790365346" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.707250 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.707691 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.207671365 +0000 UTC m=+145.291271792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.713354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" event={"ID":"e81f8e3f-96d2-4aa6-9a74-765d80da36b5","Type":"ContainerStarted","Data":"80e862bf90a117211cde670201d6489268573f149581f87bc11e5a72827d10bd"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.793473 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-2tlvs" podStartSLOduration=124.793452685 podStartE2EDuration="2m4.793452685s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.74633961 +0000 UTC m=+144.829940067" watchObservedRunningTime="2025-09-30 14:22:02.793452685 +0000 UTC m=+144.877053112" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.795078 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" event={"ID":"8a317838-165c-44ec-a74e-8fa0ebaa70d6","Type":"ContainerStarted","Data":"5707b24949e21a32a0c518d9cbb66a78110863c86c3456546a0b2d48a961c2c0"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.803396 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" podStartSLOduration=124.803373409 podStartE2EDuration="2m4.803373409s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.799754865 +0000 UTC m=+144.883355302" watchObservedRunningTime="2025-09-30 14:22:02.803373409 +0000 UTC m=+144.886973836" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.808492 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.809055 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.309037401 +0000 UTC m=+145.392637828 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.819223 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" event={"ID":"99584c4c-3c3d-4bae-99c0-2eccda0b27c0","Type":"ContainerStarted","Data":"3be99f315bf0a21fd9b26f1867d80ea3565684606de8290abc0c43a698e7a7e6"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.843120 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-w9q9n" podStartSLOduration=124.843093363 podStartE2EDuration="2m4.843093363s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.840369586 +0000 UTC m=+144.923970043" watchObservedRunningTime="2025-09-30 14:22:02.843093363 +0000 UTC m=+144.926693790" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.882861 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" event={"ID":"0095d26a-f43b-4d74-a34e-f5a3bf289dfd","Type":"ContainerStarted","Data":"4ca137528fa4d4d751e7e8a2bd3002524dbf2ecf729f0357c301f87879968b2f"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.893917 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" event={"ID":"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1","Type":"ContainerStarted","Data":"0bdc596b456e32fdab8dbab11e05be9da519517e8e463feb0117033dda69303a"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.911902 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" event={"ID":"34913ce8-3c56-4325-a430-bd2faa213576","Type":"ContainerStarted","Data":"3e70145c15a73407f41de110c256713c45cc25f74b72755d87eb0c4a4b4e0a00"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.923242 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:02 crc kubenswrapper[4799]: E0930 14:22:02.923573 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.423558592 +0000 UTC m=+145.507159019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.945764 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" event={"ID":"3449f7db-d420-4de1-9193-65dc16c8cd45","Type":"ContainerStarted","Data":"dc495a195918e37a84f7d0637d40d2676dc3ccd7bba772822b914c89cbd2e7ad"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.955481 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.960226 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jznkv" event={"ID":"2fa0f587-c724-4534-888a-33bc5f77f56a","Type":"ContainerStarted","Data":"3d7e3f1e51bea10058f7537ff27678fda03c67931f1461bf51810333559dcbd5"} Sep 30 14:22:02 crc kubenswrapper[4799]: I0930 14:22:02.983725 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" event={"ID":"7f147ef4-6fba-40b2-bdda-f19618d512eb","Type":"ContainerStarted","Data":"080b1461b14fb7b7e7794967baafbf62538826858cabdadf8a6a8811eb4272a3"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.002618 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" event={"ID":"8209103b-69e0-41be-9ce9-c87dd3e01a19","Type":"ContainerStarted","Data":"cf6b484c78974acf164a204594123403bc01feef62825e33058f36e7b34966cf"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.010991 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" event={"ID":"367e05f1-fb1d-435b-978a-83486f04ada1","Type":"ContainerStarted","Data":"6984e092a1ea7d956846984957989cb61ae1547a2d0bbdf30b7f6896d33094e5"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.028585 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.029214 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.529190539 +0000 UTC m=+145.612791006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.033615 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" event={"ID":"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd","Type":"ContainerStarted","Data":"8825ef73d3b0f29c2278b96c6dc0bed273a89c90152cff23c5ebb64222ff7f29"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.039918 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" event={"ID":"7c44311c-64e8-4587-95c6-2bdda1c93ad1","Type":"ContainerStarted","Data":"a24e7a727572fe4d5905d8d7defa0582ca073e676cf7af1d96e1ec351eb63723"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.039973 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" event={"ID":"7c44311c-64e8-4587-95c6-2bdda1c93ad1","Type":"ContainerStarted","Data":"9f219286a42a33363323e71df4d410e5b48f1d493a923b26b33a80a6a5da9641"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.060893 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-glzch" event={"ID":"6e587e9f-8d37-46ef-ab7e-23e5f772078b","Type":"ContainerStarted","Data":"c4ccea67ba2716660557a7138803c2d749fed9614dd5767bc203ade9a53786cd"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.097243 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" event={"ID":"018cb267-3656-4202-ba7e-323811533a19","Type":"ContainerStarted","Data":"ddead0e20392b2c8840f1705fca29fbdf2da56ef3d67ff309f1b866409bdfae6"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.111593 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" event={"ID":"d931fb63-da78-4311-ba16-e5b931a12dad","Type":"ContainerStarted","Data":"e85e2d6d4f2bebf43d2e4ed08437bace5fc171794388e6f05da2c1bf008c189b"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.113111 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6hvgg" podStartSLOduration=125.113093166 podStartE2EDuration="2m5.113093166s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:02.921315388 +0000 UTC m=+145.004915815" watchObservedRunningTime="2025-09-30 14:22:03.113093166 +0000 UTC m=+145.196693593" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.129498 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.129642 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.629617288 +0000 UTC m=+145.713217725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.129965 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.132966 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.632955493 +0000 UTC m=+145.716555910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.156440 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-m4jz9" event={"ID":"292776df-8fba-4f59-bcfa-e48d6bb5626d","Type":"ContainerStarted","Data":"bda53b814f119fd4a6bb5950c8d5f0de6a3e8136ed37a99349cd44fe70797c02"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.175756 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" podStartSLOduration=125.175733715 podStartE2EDuration="2m5.175733715s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:03.172251746 +0000 UTC m=+145.255852173" watchObservedRunningTime="2025-09-30 14:22:03.175733715 +0000 UTC m=+145.259334142" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.178473 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-svk57" event={"ID":"03c9ee3f-722e-4db8-a8f5-dc359486e4f4","Type":"ContainerStarted","Data":"0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.197193 4799 generic.go:334] "Generic (PLEG): container finished" podID="dedc044b-a09f-436a-8bdb-3b606269e3c5" containerID="f0d06f7944055ca580a2f59c10982507f276cef1bfc877caf5f7428d3aa043df" exitCode=0 Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.197326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" event={"ID":"dedc044b-a09f-436a-8bdb-3b606269e3c5","Type":"ContainerDied","Data":"f0d06f7944055ca580a2f59c10982507f276cef1bfc877caf5f7428d3aa043df"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.217196 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" event={"ID":"8485bb78-9aed-4175-8a10-498394ad1524","Type":"ContainerStarted","Data":"9e17359101a016fff49156c5a7d9b9cdecb7b1d321d65c8464864726ee8ee796"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.218357 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" event={"ID":"e7c75754-cd2c-436f-8eea-08e68904923b","Type":"ContainerStarted","Data":"705403a2c109f7f2b930f4649bcda65dd5e5c1e18f7f8f191ef4c27dc0cf3857"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.220048 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vxm25" event={"ID":"ae4eb023-1b23-4106-9203-9326737be923","Type":"ContainerStarted","Data":"799c0f42036d8d085b9b9c029c7eaab85fe60f94d7c86e53eb1c4952d732f1e9"} Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.225370 4799 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bfxv2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.225418 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" podUID="c35fadda-30b5-438e-b0a6-945004f36333" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.225573 4799 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-8mrvt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.225631 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" podUID="c162443e-089d-49e2-968d-c712bf5f5aba" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.225717 4799 patch_prober.go:28] interesting pod/console-operator-58897d9998-xvd7z container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/readyz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.225732 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" podUID="56b3121f-1c64-40d2-9cfd-b2c4a55451cc" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/readyz\": dial tcp 10.217.0.5:8443: connect: connection refused" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.231042 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.231755 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.731740225 +0000 UTC m=+145.815340642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.249525 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-glzch" podStartSLOduration=125.249492932 podStartE2EDuration="2m5.249492932s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:03.249026139 +0000 UTC m=+145.332626566" watchObservedRunningTime="2025-09-30 14:22:03.249492932 +0000 UTC m=+145.333093359" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.277725 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.282086 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.282155 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.309703 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-jgd4t" podStartSLOduration=125.309682561 podStartE2EDuration="2m5.309682561s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:03.309187417 +0000 UTC m=+145.392787844" watchObservedRunningTime="2025-09-30 14:22:03.309682561 +0000 UTC m=+145.393282988" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.336710 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.340260 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.840247475 +0000 UTC m=+145.923847902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.348353 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5jqc" podStartSLOduration=125.348334356 podStartE2EDuration="2m5.348334356s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:03.344047083 +0000 UTC m=+145.427647520" watchObservedRunningTime="2025-09-30 14:22:03.348334356 +0000 UTC m=+145.431934783" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.442198 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.442542 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:03.942527406 +0000 UTC m=+146.026127833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.458205 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-m4jz9" podStartSLOduration=6.458185563 podStartE2EDuration="6.458185563s" podCreationTimestamp="2025-09-30 14:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:03.408284598 +0000 UTC m=+145.491885025" watchObservedRunningTime="2025-09-30 14:22:03.458185563 +0000 UTC m=+145.541785990" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.543626 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.544068 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.044054236 +0000 UTC m=+146.127654663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.548506 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bdngs" podStartSLOduration=125.548494263 podStartE2EDuration="2m5.548494263s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:03.518449475 +0000 UTC m=+145.602049912" watchObservedRunningTime="2025-09-30 14:22:03.548494263 +0000 UTC m=+145.632094690" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.550253 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-svk57" podStartSLOduration=125.550247183 podStartE2EDuration="2m5.550247183s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:03.548205025 +0000 UTC m=+145.631805462" watchObservedRunningTime="2025-09-30 14:22:03.550247183 +0000 UTC m=+145.633847610" Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.645359 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.646012 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.145996308 +0000 UTC m=+146.229596725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.747503 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.747921 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.247904359 +0000 UTC m=+146.331504786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.848152 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.848530 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.348515923 +0000 UTC m=+146.432116350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:03 crc kubenswrapper[4799]: I0930 14:22:03.949342 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:03 crc kubenswrapper[4799]: E0930 14:22:03.949685 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.449669763 +0000 UTC m=+146.533270190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.050390 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.050605 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.550583785 +0000 UTC m=+146.634184212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.051066 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.051598 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.551573954 +0000 UTC m=+146.635174381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.152028 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.152238 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.652191258 +0000 UTC m=+146.735791685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.152322 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.152787 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.652776134 +0000 UTC m=+146.736376731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.253162 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.254034 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.753980165 +0000 UTC m=+146.837580592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.254227 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.254641 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.754630394 +0000 UTC m=+146.838230821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.268756 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" event={"ID":"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1","Type":"ContainerStarted","Data":"e87e284498ce63b636f2ccb2b7d0645bd7d71d81e90382cefcf08cf0ed33a0b4"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.279450 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" event={"ID":"8a317838-165c-44ec-a74e-8fa0ebaa70d6","Type":"ContainerStarted","Data":"ec79ddf2e806bdb609c875f4bf929f01bd7648a88d3c7827ada03e06f9ca9a91"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.281339 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.281372 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.282311 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" event={"ID":"04c8f827-2fda-4565-9186-b78529a4b427","Type":"ContainerStarted","Data":"64d86e84d70d5697e60e7e375354bea4092e151da6c394890524e8c3007d9a59"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.283326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" event={"ID":"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64","Type":"ContainerStarted","Data":"bf047ee828b3220830ca5ef081e2cd67461d40c3b766c58a7882de85011e736a"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.289251 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" event={"ID":"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd","Type":"ContainerStarted","Data":"1297a8efb779e2b6a9969b9b65bf94b81fc99acaa86acbcab71316f4bccde0af"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.298772 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" event={"ID":"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea","Type":"ContainerStarted","Data":"d76683a939c5e4053bf7f77b0d7a7609254c6ddc3a1903bee8a0cdec3048a977"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.304969 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" event={"ID":"3449f7db-d420-4de1-9193-65dc16c8cd45","Type":"ContainerStarted","Data":"0d02a49ace36e34478a7a77c714707889415ca0edf5828ab06a6599c1120aa27"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.307240 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" event={"ID":"34913ce8-3c56-4325-a430-bd2faa213576","Type":"ContainerStarted","Data":"b9ae10475b920e9ad89f643a63b5ddbee7f342fbfa0dbcabf6dc8613965aec2e"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.307975 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.312726 4799 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jj9bm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.312782 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" podUID="34913ce8-3c56-4325-a430-bd2faa213576" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.314322 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" event={"ID":"8209103b-69e0-41be-9ce9-c87dd3e01a19","Type":"ContainerStarted","Data":"7b37eb9d6a2c2575d97fbddfb804492e0c04894bfbb846f81c87fd3a28b2b7a4"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.333349 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" event={"ID":"8485bb78-9aed-4175-8a10-498394ad1524","Type":"ContainerStarted","Data":"151d52aeea4ceb99c810e9e6dca422d0545a9dec0a58602d08aec337d1fdee4a"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.343771 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" podStartSLOduration=125.34374708 podStartE2EDuration="2m5.34374708s" podCreationTimestamp="2025-09-30 14:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:04.339069476 +0000 UTC m=+146.422669903" watchObservedRunningTime="2025-09-30 14:22:04.34374708 +0000 UTC m=+146.427347507" Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.356103 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.356254 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.856233396 +0000 UTC m=+146.939833823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.356858 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.357322 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.857310347 +0000 UTC m=+146.940910774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.360880 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" event={"ID":"6fd8f596-14c4-4269-b755-b28a70c69647","Type":"ContainerStarted","Data":"173e88149b570bbe709473d4eecad8de352f7ee66559c043b166bbbd8102bb7c"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.371106 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" event={"ID":"99ec40a6-39fc-44cb-bcc6-39bd7e7019b8","Type":"ContainerStarted","Data":"5a2d49e36a73ee2ddf4d4e35548ab0dd31c9584c0324c61cab24a5268bb8c0ac"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.377741 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" event={"ID":"7f147ef4-6fba-40b2-bdda-f19618d512eb","Type":"ContainerStarted","Data":"fff6ff50d0661449750a3403f5d699e084f7b63e133787d92d17e96c512110dd"} Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.378400 4799 patch_prober.go:28] interesting pod/console-operator-58897d9998-xvd7z container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/readyz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.378454 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" podUID="56b3121f-1c64-40d2-9cfd-b2c4a55451cc" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/readyz\": dial tcp 10.217.0.5:8443: connect: connection refused" Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.378491 4799 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fzmhc container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.378552 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" podUID="2e235578-9f56-4b94-aa1b-ad79bf0f8957" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.416411 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-df8nt" podStartSLOduration=126.416382324 podStartE2EDuration="2m6.416382324s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:04.415047176 +0000 UTC m=+146.498647603" watchObservedRunningTime="2025-09-30 14:22:04.416382324 +0000 UTC m=+146.499982751" Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.449350 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" podStartSLOduration=126.449323195 podStartE2EDuration="2m6.449323195s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:04.44492485 +0000 UTC m=+146.528525277" watchObservedRunningTime="2025-09-30 14:22:04.449323195 +0000 UTC m=+146.532923612" Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.458357 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.458851 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.958831697 +0000 UTC m=+147.042432124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.462826 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.463201 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:04.963191511 +0000 UTC m=+147.046791938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.577763 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.579034 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.0790138 +0000 UTC m=+147.162614227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.679854 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.680318 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.180298073 +0000 UTC m=+147.263898500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.781380 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.28135668 +0000 UTC m=+147.364957107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.781426 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.782002 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.782471 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.282460951 +0000 UTC m=+147.366061378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.891542 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.892555 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.392531015 +0000 UTC m=+147.476131442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:04 crc kubenswrapper[4799]: I0930 14:22:04.993777 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:04 crc kubenswrapper[4799]: E0930 14:22:04.994325 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.494299302 +0000 UTC m=+147.577899729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.094698 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.094938 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.594902435 +0000 UTC m=+147.678502862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.095128 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.095531 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.595516503 +0000 UTC m=+147.679116930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.206204 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.206437 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.706406901 +0000 UTC m=+147.790007328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.206910 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.207364 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.707346838 +0000 UTC m=+147.790947265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.288943 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:05 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:05 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:05 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.289013 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.307470 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.307748 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.807703334 +0000 UTC m=+147.891303931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.385457 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" event={"ID":"04c8f827-2fda-4565-9186-b78529a4b427","Type":"ContainerStarted","Data":"83b781c41553ba189370173de2f5f1efef8feec266745d2f5866d2ed425f2a5e"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.389599 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vxm25" event={"ID":"ae4eb023-1b23-4106-9203-9326737be923","Type":"ContainerStarted","Data":"12f367c33ac3766e833c89731be559348be303c3893fb8125ecc45781b795003"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.389659 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vxm25" event={"ID":"ae4eb023-1b23-4106-9203-9326737be923","Type":"ContainerStarted","Data":"c2b1c7d775c715fd8f8a823087bf2f2c4cddf1d72e97a4c22241eccf526acbb0"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.392972 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" event={"ID":"7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea","Type":"ContainerStarted","Data":"b7f6e307ba75f99b3e62cbd443484f0d9174717084fbf183d3dcd02d913c87ec"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.399403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" event={"ID":"3449f7db-d420-4de1-9193-65dc16c8cd45","Type":"ContainerStarted","Data":"8cfa4c89ed34e0d4fdf83eb697bcd37adb8136b2d5da2b9b2b178a1ba5c45e88"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.399703 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.401171 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jznkv" event={"ID":"2fa0f587-c724-4534-888a-33bc5f77f56a","Type":"ContainerStarted","Data":"19a68f98624362079fec7109c6d6b8d9d84aa0277ea11e6d04d96fd57d217eb0"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.403721 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" event={"ID":"6fd8f596-14c4-4269-b755-b28a70c69647","Type":"ContainerStarted","Data":"8cc1a06ae6879e7642ecd32f835e13a0d3109b36aba5d0c3703bfae2f0fa6f22"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.405818 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" event={"ID":"26b28ca2-e7b1-4e6e-ba95-ac537580f7c1","Type":"ContainerStarted","Data":"a53460f861ea117eeb56947fcb7adc6a0ebeff5c46eb7b141e4bae0c8cfd27d8"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.408879 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" event={"ID":"367e05f1-fb1d-435b-978a-83486f04ada1","Type":"ContainerStarted","Data":"d14a0fd5fbb3a5df8bd47e35d14f86259355de9e16ac97ee9d3e15cbd24291a2"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.410199 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.411625 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:05.911608852 +0000 UTC m=+147.995209279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.417094 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" event={"ID":"dedc044b-a09f-436a-8bdb-3b606269e3c5","Type":"ContainerStarted","Data":"40f2dd9cb07d3db15796c4a609532e9a06b4c918b35877689d65d71a9bec98ae"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.417920 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.423890 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" event={"ID":"99584c4c-3c3d-4bae-99c0-2eccda0b27c0","Type":"ContainerStarted","Data":"021cdc333842418ed1e248cd2abe5e59d2572a9ad3f6a865b8f29ff3a9f54f77"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.428220 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" event={"ID":"d8918fed-2347-4360-a4e1-d84658a2211f","Type":"ContainerStarted","Data":"b7a6b2aa2e5bf03f6e117ef2e18ad99045b237bcddfe99114e27c690acdbc86e"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.434666 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" event={"ID":"d931fb63-da78-4311-ba16-e5b931a12dad","Type":"ContainerStarted","Data":"1b302224f8cec8925a047479eec46b541368d7a29295313e223a5484e6345c92"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.440412 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" event={"ID":"99ec40a6-39fc-44cb-bcc6-39bd7e7019b8","Type":"ContainerStarted","Data":"be257637f35966933e62e3c60d2089c7bcfa7e775c137a6ceb446725428c3299"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.440468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" event={"ID":"99ec40a6-39fc-44cb-bcc6-39bd7e7019b8","Type":"ContainerStarted","Data":"e5c3addacaa97f62486e9b57616cc5c7c94fee4b7c1afae5eee8d51daef18bbc"} Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.444111 4799 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fzmhc container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.444159 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" podUID="2e235578-9f56-4b94-aa1b-ad79bf0f8957" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.444509 4799 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jj9bm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.444532 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" podUID="34913ce8-3c56-4325-a430-bd2faa213576" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.459745 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-dzk4q" podStartSLOduration=127.459703666 podStartE2EDuration="2m7.459703666s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.457598786 +0000 UTC m=+147.541199203" watchObservedRunningTime="2025-09-30 14:22:05.459703666 +0000 UTC m=+147.543304093" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.514568 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.514933 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.014889523 +0000 UTC m=+148.098489950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.578484 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z7hn5" podStartSLOduration=126.578454008 podStartE2EDuration="2m6.578454008s" podCreationTimestamp="2025-09-30 14:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.511176227 +0000 UTC m=+147.594776654" watchObservedRunningTime="2025-09-30 14:22:05.578454008 +0000 UTC m=+147.662054435" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.579288 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" podStartSLOduration=127.579283342 podStartE2EDuration="2m7.579283342s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.56099728 +0000 UTC m=+147.644597707" watchObservedRunningTime="2025-09-30 14:22:05.579283342 +0000 UTC m=+147.662883769" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.623958 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.625286 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.125259665 +0000 UTC m=+148.208860292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.636944 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cvsnb" podStartSLOduration=127.636911008 podStartE2EDuration="2m7.636911008s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.609404723 +0000 UTC m=+147.693005160" watchObservedRunningTime="2025-09-30 14:22:05.636911008 +0000 UTC m=+147.720511435" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.726001 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.726218 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.226186778 +0000 UTC m=+148.309787205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.726354 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.726795 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.226783226 +0000 UTC m=+148.310383823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.738881 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xgs5n" podStartSLOduration=127.7388595 podStartE2EDuration="2m7.7388595s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.737644286 +0000 UTC m=+147.821244733" watchObservedRunningTime="2025-09-30 14:22:05.7388595 +0000 UTC m=+147.822459937" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.740942 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-hrm8v" podStartSLOduration=126.74093552 podStartE2EDuration="2m6.74093552s" podCreationTimestamp="2025-09-30 14:19:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.667309497 +0000 UTC m=+147.750909934" watchObservedRunningTime="2025-09-30 14:22:05.74093552 +0000 UTC m=+147.824535947" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.777156 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" podStartSLOduration=127.777103033 podStartE2EDuration="2m7.777103033s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.776033352 +0000 UTC m=+147.859633779" watchObservedRunningTime="2025-09-30 14:22:05.777103033 +0000 UTC m=+147.860703490" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.818989 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jznkv" podStartSLOduration=8.818940258 podStartE2EDuration="8.818940258s" podCreationTimestamp="2025-09-30 14:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.815267153 +0000 UTC m=+147.898867590" watchObservedRunningTime="2025-09-30 14:22:05.818940258 +0000 UTC m=+147.902540685" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.827728 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.828310 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.328293945 +0000 UTC m=+148.411894372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.890843 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xhs4m" podStartSLOduration=127.890801781 podStartE2EDuration="2m7.890801781s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.888543436 +0000 UTC m=+147.972143863" watchObservedRunningTime="2025-09-30 14:22:05.890801781 +0000 UTC m=+147.974402208" Sep 30 14:22:05 crc kubenswrapper[4799]: I0930 14:22:05.929878 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:05 crc kubenswrapper[4799]: E0930 14:22:05.930416 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.430403012 +0000 UTC m=+148.514003439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.030933 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.031152 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.531111689 +0000 UTC m=+148.614712116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.031298 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.031590 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.531578082 +0000 UTC m=+148.615178509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.074855 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-p55n7" podStartSLOduration=128.074835088 podStartE2EDuration="2m8.074835088s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:05.982988094 +0000 UTC m=+148.066588521" watchObservedRunningTime="2025-09-30 14:22:06.074835088 +0000 UTC m=+148.158435515" Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.076376 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" podStartSLOduration=128.076370852 podStartE2EDuration="2m8.076370852s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:06.071528183 +0000 UTC m=+148.155128610" watchObservedRunningTime="2025-09-30 14:22:06.076370852 +0000 UTC m=+148.159971269" Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.133165 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.133431 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.633395561 +0000 UTC m=+148.716995988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.133709 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.134094 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.63408357 +0000 UTC m=+148.717683997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.227923 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" podStartSLOduration=128.2278973 podStartE2EDuration="2m8.2278973s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:06.116406025 +0000 UTC m=+148.200006462" watchObservedRunningTime="2025-09-30 14:22:06.2278973 +0000 UTC m=+148.311497727" Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.235641 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.235926 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.735891638 +0000 UTC m=+148.819492065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.236170 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.236606 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.736589828 +0000 UTC m=+148.820190255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.280586 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:06 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:06 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:06 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.280715 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.337955 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.338217 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.8381842 +0000 UTC m=+148.921784627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.339864 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.340421 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.840406774 +0000 UTC m=+148.924007201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.345366 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" podStartSLOduration=128.345341335 podStartE2EDuration="2m8.345341335s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:06.235538688 +0000 UTC m=+148.319139135" watchObservedRunningTime="2025-09-30 14:22:06.345341335 +0000 UTC m=+148.428941782" Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.441805 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.442041 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.942022237 +0000 UTC m=+149.025622664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.442335 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.442819 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:06.942796989 +0000 UTC m=+149.026397416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.450286 4799 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jj9bm container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.450359 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" podUID="34913ce8-3c56-4325-a430-bd2faa213576" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.452620 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-8wmvv" podStartSLOduration=128.452609349 podStartE2EDuration="2m8.452609349s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:06.347859117 +0000 UTC m=+148.431459544" watchObservedRunningTime="2025-09-30 14:22:06.452609349 +0000 UTC m=+148.536209766" Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.456005 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cp7sz" podStartSLOduration=128.455994326 podStartE2EDuration="2m8.455994326s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:06.446386101 +0000 UTC m=+148.529986538" watchObservedRunningTime="2025-09-30 14:22:06.455994326 +0000 UTC m=+148.539594753" Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.543544 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.543753 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.043724762 +0000 UTC m=+149.127325189 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.546428 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.548822 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.048799067 +0000 UTC m=+149.132399674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.647942 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.648358 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.14833889 +0000 UTC m=+149.231939317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.750450 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.751297 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.25124947 +0000 UTC m=+149.334849897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.852280 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.852424 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.352397639 +0000 UTC m=+149.435998056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.852534 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.852918 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.352903043 +0000 UTC m=+149.436503470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:06 crc kubenswrapper[4799]: I0930 14:22:06.954167 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:06 crc kubenswrapper[4799]: E0930 14:22:06.954808 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.454787884 +0000 UTC m=+149.538388311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.055996 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.056624 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.556594322 +0000 UTC m=+149.640194749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.157862 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.158148 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.658107732 +0000 UTC m=+149.741708159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.158274 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.158847 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.658838953 +0000 UTC m=+149.742439380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.259518 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.259733 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.759705024 +0000 UTC m=+149.843305451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.259923 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.260573 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.760540698 +0000 UTC m=+149.844141125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.280771 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:07 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:07 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:07 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.280838 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.361541 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.861525072 +0000 UTC m=+149.945125499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.361422 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.361936 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.362237 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.862228343 +0000 UTC m=+149.945828770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.452174 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" event={"ID":"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64","Type":"ContainerStarted","Data":"60f4c1467589c29dc4601337ccff9917dfc539664f8cfca3d0b4bc746edfe6a9"} Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.463065 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.463480 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.464629 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:07.964610537 +0000 UTC m=+150.048210964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.481157 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.565449 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.565516 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.566829 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.566982 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.567424 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.067406094 +0000 UTC m=+150.151006521 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.569525 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.570169 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.607364 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.668280 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.668601 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.168566383 +0000 UTC m=+150.252166850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.668752 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.669138 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.169127889 +0000 UTC m=+150.252728316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.715381 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.769358 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.769563 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.269503937 +0000 UTC m=+150.353104364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.769610 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.770043 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.270034912 +0000 UTC m=+150.353635409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.825541 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.832590 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.871354 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.871562 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.371529731 +0000 UTC m=+150.455130158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.871623 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.872138 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.372122058 +0000 UTC m=+150.455722485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:07 crc kubenswrapper[4799]: I0930 14:22:07.972791 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:07 crc kubenswrapper[4799]: E0930 14:22:07.973082 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.473066861 +0000 UTC m=+150.556667288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.081533 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.082030 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.582014844 +0000 UTC m=+150.665615271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.182441 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.183133 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.683117792 +0000 UTC m=+150.766718219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.280969 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:08 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:08 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:08 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.281072 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.285087 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.285581 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.785559268 +0000 UTC m=+150.869159705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.387168 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.387390 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.887361826 +0000 UTC m=+150.970962253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.387441 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.387794 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.887782098 +0000 UTC m=+150.971382525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.467510 4799 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-mgzkz container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.467559 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" podUID="dedc044b-a09f-436a-8bdb-3b606269e3c5" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.467800 4799 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-mgzkz container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.467816 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" podUID="dedc044b-a09f-436a-8bdb-3b606269e3c5" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.488387 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.488737 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:08.98872188 +0000 UTC m=+151.072322307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.527332 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vxm25" podStartSLOduration=11.527310113 podStartE2EDuration="11.527310113s" podCreationTimestamp="2025-09-30 14:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:06.560585143 +0000 UTC m=+148.644185570" watchObservedRunningTime="2025-09-30 14:22:08.527310113 +0000 UTC m=+150.610910540" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.527931 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-76bw2"] Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.530398 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.559104 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.591736 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76bw2"] Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.593718 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5f8w\" (UniqueName: \"kubernetes.io/projected/4eb004e8-ef9d-404c-9321-2c566b86e393-kube-api-access-z5f8w\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.593772 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.593797 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-catalog-content\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.593859 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-utilities\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.594140 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.094128481 +0000 UTC m=+151.177728908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.611151 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.685928 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.696102 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.696389 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-utilities\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.696462 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5f8w\" (UniqueName: \"kubernetes.io/projected/4eb004e8-ef9d-404c-9321-2c566b86e393-kube-api-access-z5f8w\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.696498 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-catalog-content\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.697404 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.197379431 +0000 UTC m=+151.280979848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.697865 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-utilities\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.698150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-catalog-content\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.791727 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.792686 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.808356 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.808950 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.308932547 +0000 UTC m=+151.392532974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: W0930 14:22:08.814039 4799 reflector.go:561] object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n": failed to list *v1.Secret: secrets "installer-sa-dockercfg-kjl2n" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-kube-controller-manager": no relationship found between node 'crc' and this object Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.814095 4799 reflector.go:158] "Unhandled Error" err="object-\"openshift-kube-controller-manager\"/\"installer-sa-dockercfg-kjl2n\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"installer-sa-dockercfg-kjl2n\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-kube-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 30 14:22:08 crc kubenswrapper[4799]: W0930 14:22:08.814188 4799 reflector.go:561] object-"openshift-kube-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-kube-controller-manager": no relationship found between node 'crc' and this object Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.814202 4799 reflector.go:158] "Unhandled Error" err="object-\"openshift-kube-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-kube-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.914919 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.914956 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.915186 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.915269 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:08 crc kubenswrapper[4799]: E0930 14:22:08.915482 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.4154615 +0000 UTC m=+151.499061927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.915478 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.919722 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zrqpk"] Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.920567 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.940057 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.959181 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 14:22:08 crc kubenswrapper[4799]: I0930 14:22:08.974597 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.007208 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5f8w\" (UniqueName: \"kubernetes.io/projected/4eb004e8-ef9d-404c-9321-2c566b86e393-kube-api-access-z5f8w\") pod \"community-operators-76bw2\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.018973 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-catalog-content\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.019145 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.019264 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.019291 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.019316 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-utilities\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.019355 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blpcv\" (UniqueName: \"kubernetes.io/projected/70b0c843-7e0f-4cca-8443-9306f0c7ad76-kube-api-access-blpcv\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.023853 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.023929 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.024279 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.024325 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.024719 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.024738 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.024922 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.025085 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.525072812 +0000 UTC m=+151.608673239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.057330 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nzj2q"] Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.058580 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.123267 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.123693 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-catalog-content\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.123731 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnslj\" (UniqueName: \"kubernetes.io/projected/674513e3-aa0c-4814-884c-d0cb1468a3c9-kube-api-access-jnslj\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.123766 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-catalog-content\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.123799 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-utilities\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.123871 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-utilities\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.123901 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blpcv\" (UniqueName: \"kubernetes.io/projected/70b0c843-7e0f-4cca-8443-9306f0c7ad76-kube-api-access-blpcv\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.124144 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.624116271 +0000 UTC m=+151.707716698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.124712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-catalog-content\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.124978 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-utilities\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.139488 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.144722 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nzj2q"] Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.162210 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrqpk"] Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.176178 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9vmjm"] Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.177151 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.208846 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.209905 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.214944 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blpcv\" (UniqueName: \"kubernetes.io/projected/70b0c843-7e0f-4cca-8443-9306f0c7ad76-kube-api-access-blpcv\") pod \"community-operators-zrqpk\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.216920 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.225711 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnslj\" (UniqueName: \"kubernetes.io/projected/674513e3-aa0c-4814-884c-d0cb1468a3c9-kube-api-access-jnslj\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.225796 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7c6f\" (UniqueName: \"kubernetes.io/projected/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-kube-api-access-h7c6f\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.225853 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-catalog-content\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.225945 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-utilities\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.226593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-utilities\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.226519 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-catalog-content\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.226741 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.226812 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-catalog-content\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.227071 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-utilities\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.227243 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.727231876 +0000 UTC m=+151.810832303 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.279815 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.284643 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnslj\" (UniqueName: \"kubernetes.io/projected/674513e3-aa0c-4814-884c-d0cb1468a3c9-kube-api-access-jnslj\") pod \"certified-operators-nzj2q\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.300247 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9vmjm"] Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.314232 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:09 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:09 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:09 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.319832 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.336912 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.337818 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-utilities\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.338130 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-catalog-content\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.338237 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7c6f\" (UniqueName: \"kubernetes.io/projected/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-kube-api-access-h7c6f\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.339029 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.839012839 +0000 UTC m=+151.922613266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.340338 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-utilities\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.341561 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-catalog-content\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.395156 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-8mrvt" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.409685 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.410895 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7c6f\" (UniqueName: \"kubernetes.io/projected/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-kube-api-access-h7c6f\") pod \"certified-operators-9vmjm\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.416028 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vxm25" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.464284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.469537 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:09.969520767 +0000 UTC m=+152.053121194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.515076 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-xvd7z" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.585211 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.586120 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.590735 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.090712749 +0000 UTC m=+152.174313166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.667039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"eb936d986255c80bd2f7ce62371aac5a884fe67912a8e672bbb70f31f487c376"} Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.689822 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.690674 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.190642294 +0000 UTC m=+152.274242721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.715275 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"03aa7c0417bf11f46efd8a0e6012d55323c1f6149d92834bf85d5a05bd41a13e"} Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.750841 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.785881 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.796788 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.797497 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.297483686 +0000 UTC m=+152.381084113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:09 crc kubenswrapper[4799]: I0930 14:22:09.900400 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:09 crc kubenswrapper[4799]: E0930 14:22:09.900730 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.400718475 +0000 UTC m=+152.484318902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.001166 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.001628 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.501613007 +0000 UTC m=+152.585213444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.106835 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.107150 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.607138011 +0000 UTC m=+152.690738438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.178951 4799 projected.go:288] Couldn't get configMap openshift-kube-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.179012 4799 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-controller-manager/revision-pruner-9-crc: failed to sync configmap cache: timed out waiting for the condition Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.179078 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access podName:bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0 nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.679060346 +0000 UTC m=+152.762660773 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access") pod "revision-pruner-9-crc" (UID: "bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0") : failed to sync configmap cache: timed out waiting for the condition Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.188595 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.188638 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.189823 4799 patch_prober.go:28] interesting pod/console-f9d7485db-svk57 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.189864 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-svk57" podUID="03c9ee3f-722e-4db8-a8f5-dc359486e4f4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.207816 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.207924 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.70790505 +0000 UTC m=+152.791505477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.208406 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.208697 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.708688492 +0000 UTC m=+152.792288919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.260825 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzmhc" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.283836 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.287023 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:10 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:10 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:10 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.287089 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.309435 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.310779 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.810759398 +0000 UTC m=+152.894359825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.311305 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.311669 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.811641223 +0000 UTC m=+152.895241650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.329285 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.347414 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.348313 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-b7pv9 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.348345 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" podUID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.348486 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-b7pv9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.348558 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" podUID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.349009 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-b7pv9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.349029 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" podUID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.415042 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.417826 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:10.917810216 +0000 UTC m=+153.001410643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.518191 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.518862 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.018850612 +0000 UTC m=+153.102451039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.621031 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.628369 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.12834155 +0000 UTC m=+153.211941987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.632708 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.633153 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.133135717 +0000 UTC m=+153.216736144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.699205 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jj9bm" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.734121 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.734381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.735858 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.23583689 +0000 UTC m=+153.319437317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.764614 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.782349 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"bff446e038b581c3ac22e177aab77454afbcef32d603e85576a548bceb9b9c5d"} Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.782401 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"74eda2eb5cd4dd7289788cdccd3b1ab12ac08902df9d4a973d8ee59538d414c9"} Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.784758 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.825479 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" event={"ID":"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64","Type":"ContainerStarted","Data":"56b541028ba688e8a7da936132b44b62b2655e172a8aa4aedc89da0501fdf406"} Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.842948 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.847310 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.347288704 +0000 UTC m=+153.430889131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.847516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7064fdb8354283573d65aca9f41226b32966c838bcc161532678c8848fdbd9ba"} Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.876679 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"bf48399e76dc34f160a35e3424894b4f22c69d3ca83f90d0b9213f44c0d87dc2"} Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.894128 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x52bz" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.899293 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pzj2v"] Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.900569 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.902451 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.934920 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.944301 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:10 crc kubenswrapper[4799]: E0930 14:22:10.945596 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.445576171 +0000 UTC m=+153.529176598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:10 crc kubenswrapper[4799]: I0930 14:22:10.953482 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pzj2v"] Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.051430 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.051484 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-catalog-content\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.051500 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgh5n\" (UniqueName: \"kubernetes.io/projected/fe8c96af-648f-407f-a61b-d09601743327-kube-api-access-lgh5n\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.051557 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-utilities\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.053355 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.55334192 +0000 UTC m=+153.636942347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.154783 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.154942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-utilities\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.155040 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-catalog-content\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.155067 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgh5n\" (UniqueName: \"kubernetes.io/projected/fe8c96af-648f-407f-a61b-d09601743327-kube-api-access-lgh5n\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.155452 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.655434176 +0000 UTC m=+153.739034603 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.155828 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-utilities\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.156056 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-catalog-content\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.250182 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgh5n\" (UniqueName: \"kubernetes.io/projected/fe8c96af-648f-407f-a61b-d09601743327-kube-api-access-lgh5n\") pod \"redhat-marketplace-pzj2v\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.256367 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.256670 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.756641097 +0000 UTC m=+153.840241524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.287942 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:11 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:11 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:11 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.287996 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.323451 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.358069 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.358406 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.858375053 +0000 UTC m=+153.941975480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.457562 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nh8v4"] Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.459483 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.459826 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:11.959811511 +0000 UTC m=+154.043411938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.459958 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.472106 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mgzkz" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.560723 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.560970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-catalog-content\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.561075 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.061054623 +0000 UTC m=+154.144655050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.561098 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8flc\" (UniqueName: \"kubernetes.io/projected/5ef93211-55a5-43d0-9bda-a1e41c44fb35-kube-api-access-d8flc\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.561175 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-utilities\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.572179 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh8v4"] Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.643440 4799 patch_prober.go:28] interesting pod/apiserver-76f77b778f-clzwz container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]log ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]etcd ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/start-apiserver-admission-initializer ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/generic-apiserver-start-informers ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/max-in-flight-filter ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/storage-object-count-tracker-hook ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/image.openshift.io-apiserver-caches ok Sep 30 14:22:11 crc kubenswrapper[4799]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Sep 30 14:22:11 crc kubenswrapper[4799]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/project.openshift.io-projectcache ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Sep 30 14:22:11 crc kubenswrapper[4799]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/openshift.io-restmapperupdater ok Sep 30 14:22:11 crc kubenswrapper[4799]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Sep 30 14:22:11 crc kubenswrapper[4799]: livez check failed Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.643509 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" podUID="7a56ad12-0b83-4023-b2d1-b3b6f9e5f1ea" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.664004 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8flc\" (UniqueName: \"kubernetes.io/projected/5ef93211-55a5-43d0-9bda-a1e41c44fb35-kube-api-access-d8flc\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.664098 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-utilities\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.664122 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-catalog-content\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.664154 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.665702 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.165686022 +0000 UTC m=+154.249286449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.665776 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-catalog-content\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.666127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-utilities\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.765249 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.765412 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.26538687 +0000 UTC m=+154.348987307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.765558 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.765907 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.265895804 +0000 UTC m=+154.349496231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.867131 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.867491 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.367475466 +0000 UTC m=+154.451075893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.874189 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8flc\" (UniqueName: \"kubernetes.io/projected/5ef93211-55a5-43d0-9bda-a1e41c44fb35-kube-api-access-d8flc\") pod \"redhat-marketplace-nh8v4\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.887399 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76bw2"] Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.902719 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" event={"ID":"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64","Type":"ContainerStarted","Data":"28f197f395b4b0228634ffb438bc9844631a08c22adce7cb9d1ca7a6536f3388"} Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.925904 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ntf96"] Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.927484 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.952074 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 30 14:22:11 crc kubenswrapper[4799]: I0930 14:22:11.969800 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:11 crc kubenswrapper[4799]: E0930 14:22:11.970176 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.470156039 +0000 UTC m=+154.553756476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.054375 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ntf96"] Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.073443 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.074042 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7vqt\" (UniqueName: \"kubernetes.io/projected/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-kube-api-access-j7vqt\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.074142 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-catalog-content\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.074281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-utilities\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.075448 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.575417515 +0000 UTC m=+154.659017942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.102175 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.124036 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pdhnm"] Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.125615 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.136760 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9vmjm"] Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.181952 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7vqt\" (UniqueName: \"kubernetes.io/projected/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-kube-api-access-j7vqt\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.182013 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-catalog-content\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.182046 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-catalog-content\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.182072 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qfl7\" (UniqueName: \"kubernetes.io/projected/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-kube-api-access-4qfl7\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.182092 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-utilities\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.182116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.182137 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-utilities\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.184984 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-utilities\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.185297 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.685281973 +0000 UTC m=+154.768882400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.199096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-catalog-content\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.215717 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.216613 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.230756 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pdhnm"] Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.239239 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrqpk"] Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.272902 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.288146 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.288425 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-catalog-content\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.288463 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qfl7\" (UniqueName: \"kubernetes.io/projected/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-kube-api-access-4qfl7\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.288503 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-utilities\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.288547 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.288613 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.288773 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.788752579 +0000 UTC m=+154.872353006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.289213 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-catalog-content\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.289777 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-utilities\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.291898 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:12 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:12 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:12 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.291944 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.367137 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.370032 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nzj2q"] Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.379941 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qfl7\" (UniqueName: \"kubernetes.io/projected/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-kube-api-access-4qfl7\") pod \"redhat-operators-pdhnm\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.389293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.389376 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.389414 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.389451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.389829 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.889811376 +0000 UTC m=+154.973411883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.405737 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7vqt\" (UniqueName: \"kubernetes.io/projected/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-kube-api-access-j7vqt\") pod \"redhat-operators-ntf96\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.425184 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.486886 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.494628 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.494820 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.494903 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:12.994888787 +0000 UTC m=+155.078489214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.573936 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.599760 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.600132 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.100117773 +0000 UTC m=+155.183718200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.628857 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.703078 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.703549 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.203521466 +0000 UTC m=+155.287121883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.805274 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.305240622 +0000 UTC m=+155.388841049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.805864 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.906742 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.906867 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.406842945 +0000 UTC m=+155.490443372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.906924 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:12 crc kubenswrapper[4799]: E0930 14:22:12.907239 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.407228876 +0000 UTC m=+155.490829303 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:12 crc kubenswrapper[4799]: I0930 14:22:12.935524 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrqpk" event={"ID":"70b0c843-7e0f-4cca-8443-9306f0c7ad76","Type":"ContainerStarted","Data":"cb1e8b5dc9538e9d403d0cbc5c33d419834d80a3a7378e6b4dd4ada7f56a7ea3"} Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.000892 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vmjm" event={"ID":"5b9dfdd6-a206-4a32-8e66-4402828d4d6c","Type":"ContainerStarted","Data":"882e9d4819e4f5f61dd09803f22b1cf1a804c1a3e4cb1535f53337f9e68e662f"} Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.008847 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.009287 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.50925104 +0000 UTC m=+155.592851467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.015906 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" event={"ID":"d119c9e7-f611-40ef-8c19-fb5a1ed9cf64","Type":"ContainerStarted","Data":"bd0aa5fb2988eb47b6613681dce83fd912ed0c4d177d0eedbd1c7602d0eaf029"} Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.018245 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2q" event={"ID":"674513e3-aa0c-4814-884c-d0cb1468a3c9","Type":"ContainerStarted","Data":"af4f891e954a5b8041b1baec0b3b8d70853600f1b2733e851e39241c23b1bb4b"} Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.056218 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76bw2" event={"ID":"4eb004e8-ef9d-404c-9321-2c566b86e393","Type":"ContainerStarted","Data":"b29181a6b5ca736b39b9963fe174acf631f3cac125731f5e50548edb69f593bc"} Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.056538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76bw2" event={"ID":"4eb004e8-ef9d-404c-9321-2c566b86e393","Type":"ContainerStarted","Data":"742e75b947f7e922960ba5edcbbdf5deb50ec555354d4f52fe1ae4aef0041101"} Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.062962 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.066161 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-mkz8h" podStartSLOduration=16.066136345 podStartE2EDuration="16.066136345s" podCreationTimestamp="2025-09-30 14:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:13.05440548 +0000 UTC m=+155.138005927" watchObservedRunningTime="2025-09-30 14:22:13.066136345 +0000 UTC m=+155.149736772" Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.087643 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pzj2v"] Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.116759 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.118037 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.618025237 +0000 UTC m=+155.701625664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: W0930 14:22:13.155104 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe8c96af_648f_407f_a61b_d09601743327.slice/crio-f222c0d2e8828998b9250abca65a28f6fa91136ab1db51acd6e284936f504879 WatchSource:0}: Error finding container f222c0d2e8828998b9250abca65a28f6fa91136ab1db51acd6e284936f504879: Status 404 returned error can't find the container with id f222c0d2e8828998b9250abca65a28f6fa91136ab1db51acd6e284936f504879 Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.221981 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.222307 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.722292336 +0000 UTC m=+155.805892763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.295248 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:13 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:13 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:13 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.295520 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.323304 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.323640 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.82362693 +0000 UTC m=+155.907227357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.401542 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.431596 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.440326 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.940293183 +0000 UTC m=+156.023893610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.440432 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.449005 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:13.948988261 +0000 UTC m=+156.032588688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.464155 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh8v4"] Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.558588 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.558801 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:14.058755117 +0000 UTC m=+156.142355544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.558993 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.559400 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:14.059389805 +0000 UTC m=+156.142990232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.615473 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pdhnm"] Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.626223 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.660641 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.660899 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-30 14:22:14.160873344 +0000 UTC m=+156.244473771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.661192 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:13 crc kubenswrapper[4799]: E0930 14:22:13.661523 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-30 14:22:14.161509452 +0000 UTC m=+156.245109879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pnkgs" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.674767 4799 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Sep 30 14:22:13 crc kubenswrapper[4799]: W0930 14:22:13.676099 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9eab5bb_9ccf_4184_a82e_25b41844ea1b.slice/crio-3063ea6c02d5bce2c8bdf7498786ca50d53c797a66645c79d418fbdb90822ea2 WatchSource:0}: Error finding container 3063ea6c02d5bce2c8bdf7498786ca50d53c797a66645c79d418fbdb90822ea2: Status 404 returned error can't find the container with id 3063ea6c02d5bce2c8bdf7498786ca50d53c797a66645c79d418fbdb90822ea2 Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.696091 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ntf96"] Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.708011 4799 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-09-30T14:22:13.674789942Z","Handler":null,"Name":""} Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.744839 4799 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.744887 4799 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.762123 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.778348 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.864174 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.884009 4799 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.884046 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.911630 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.917636 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-clzwz" Sep 30 14:22:13 crc kubenswrapper[4799]: I0930 14:22:13.931948 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pnkgs\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.008902 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.073594 4799 generic.go:334] "Generic (PLEG): container finished" podID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerID="6880bf8456dfd4e03ded124a780e099294a4a83057fd6e7c678eba615f5fc265" exitCode=0 Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.074785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2q" event={"ID":"674513e3-aa0c-4814-884c-d0cb1468a3c9","Type":"ContainerDied","Data":"6880bf8456dfd4e03ded124a780e099294a4a83057fd6e7c678eba615f5fc265"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.084471 4799 generic.go:334] "Generic (PLEG): container finished" podID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerID="c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7" exitCode=0 Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.084736 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh8v4" event={"ID":"5ef93211-55a5-43d0-9bda-a1e41c44fb35","Type":"ContainerDied","Data":"c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.084785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh8v4" event={"ID":"5ef93211-55a5-43d0-9bda-a1e41c44fb35","Type":"ContainerStarted","Data":"8520605c506fde2c7c70ad2634c72024cd3ac1cce9402efad2b06147f5855f4b"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.090696 4799 generic.go:334] "Generic (PLEG): container finished" podID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerID="962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f" exitCode=0 Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.090755 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrqpk" event={"ID":"70b0c843-7e0f-4cca-8443-9306f0c7ad76","Type":"ContainerDied","Data":"962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.102762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec017ab4-5210-40d6-a33a-ca7ace14bafe","Type":"ContainerStarted","Data":"c6cf7369eb5a7c4fa2cd1f3e1f2fc48793ee3594201c41accd464dc6b11a5127"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.118858 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0","Type":"ContainerStarted","Data":"97887ffdbe31e0279232737c494a41510457cc30b75805e5b8fa4cfb520e2870"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.132369 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdhnm" event={"ID":"b9eab5bb-9ccf-4184-a82e-25b41844ea1b","Type":"ContainerStarted","Data":"4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.132407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdhnm" event={"ID":"b9eab5bb-9ccf-4184-a82e-25b41844ea1b","Type":"ContainerStarted","Data":"3063ea6c02d5bce2c8bdf7498786ca50d53c797a66645c79d418fbdb90822ea2"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.144822 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntf96" event={"ID":"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f","Type":"ContainerStarted","Data":"b66989d24033687a618f63d84920f67f78b92e148b29eb50e127aac162ccb313"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.144872 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntf96" event={"ID":"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f","Type":"ContainerStarted","Data":"3656e658885737a4776fc63482390307c4b6c0b8197c41ea78faa459ab982726"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.169632 4799 generic.go:334] "Generic (PLEG): container finished" podID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerID="b29181a6b5ca736b39b9963fe174acf631f3cac125731f5e50548edb69f593bc" exitCode=0 Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.169723 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76bw2" event={"ID":"4eb004e8-ef9d-404c-9321-2c566b86e393","Type":"ContainerDied","Data":"b29181a6b5ca736b39b9963fe174acf631f3cac125731f5e50548edb69f593bc"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.187225 4799 generic.go:334] "Generic (PLEG): container finished" podID="fe8c96af-648f-407f-a61b-d09601743327" containerID="373648fdc330acb0b2c9c2831cf7f0335c99a6d524a04d87be02842974830083" exitCode=0 Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.187392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pzj2v" event={"ID":"fe8c96af-648f-407f-a61b-d09601743327","Type":"ContainerDied","Data":"373648fdc330acb0b2c9c2831cf7f0335c99a6d524a04d87be02842974830083"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.187496 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pzj2v" event={"ID":"fe8c96af-648f-407f-a61b-d09601743327","Type":"ContainerStarted","Data":"f222c0d2e8828998b9250abca65a28f6fa91136ab1db51acd6e284936f504879"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.194309 4799 generic.go:334] "Generic (PLEG): container finished" podID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerID="65a963cbbaf84278fe18927cdadee74e0f5be417433733910c961070697ef4eb" exitCode=0 Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.195435 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vmjm" event={"ID":"5b9dfdd6-a206-4a32-8e66-4402828d4d6c","Type":"ContainerDied","Data":"65a963cbbaf84278fe18927cdadee74e0f5be417433733910c961070697ef4eb"} Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.294778 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:14 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:14 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:14 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.294832 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.528265 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Sep 30 14:22:14 crc kubenswrapper[4799]: I0930 14:22:14.558422 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pnkgs"] Sep 30 14:22:14 crc kubenswrapper[4799]: W0930 14:22:14.586713 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7675ad08_1655_418f_b0c1_baac6a2d35b6.slice/crio-9adcc946fb6d41e0e23b0b9ce071785c0ebe473eeef4817dd03f457ceea0c246 WatchSource:0}: Error finding container 9adcc946fb6d41e0e23b0b9ce071785c0ebe473eeef4817dd03f457ceea0c246: Status 404 returned error can't find the container with id 9adcc946fb6d41e0e23b0b9ce071785c0ebe473eeef4817dd03f457ceea0c246 Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.229699 4799 generic.go:334] "Generic (PLEG): container finished" podID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerID="b66989d24033687a618f63d84920f67f78b92e148b29eb50e127aac162ccb313" exitCode=0 Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.229842 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntf96" event={"ID":"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f","Type":"ContainerDied","Data":"b66989d24033687a618f63d84920f67f78b92e148b29eb50e127aac162ccb313"} Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.232162 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" event={"ID":"7675ad08-1655-418f-b0c1-baac6a2d35b6","Type":"ContainerStarted","Data":"cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28"} Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.232189 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" event={"ID":"7675ad08-1655-418f-b0c1-baac6a2d35b6","Type":"ContainerStarted","Data":"9adcc946fb6d41e0e23b0b9ce071785c0ebe473eeef4817dd03f457ceea0c246"} Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.232480 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.236316 4799 generic.go:334] "Generic (PLEG): container finished" podID="ec017ab4-5210-40d6-a33a-ca7ace14bafe" containerID="ad1be66b7e54eb7870497e05c4f578f8534d0ce30f138f7cbd44ab7a8c787802" exitCode=0 Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.236374 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec017ab4-5210-40d6-a33a-ca7ace14bafe","Type":"ContainerDied","Data":"ad1be66b7e54eb7870497e05c4f578f8534d0ce30f138f7cbd44ab7a8c787802"} Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.240664 4799 generic.go:334] "Generic (PLEG): container finished" podID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerID="4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24" exitCode=0 Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.240784 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdhnm" event={"ID":"b9eab5bb-9ccf-4184-a82e-25b41844ea1b","Type":"ContainerDied","Data":"4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24"} Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.274338 4799 generic.go:334] "Generic (PLEG): container finished" podID="bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0" containerID="0d18ad8cb4d5df1a15e3f860a1e3949e53cdbe91a4783b7db4e7f4a05898a295" exitCode=0 Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.274412 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0","Type":"ContainerDied","Data":"0d18ad8cb4d5df1a15e3f860a1e3949e53cdbe91a4783b7db4e7f4a05898a295"} Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.291553 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:15 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:15 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:15 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.291621 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.330486 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" podStartSLOduration=137.330468127 podStartE2EDuration="2m17.330468127s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:15.330098666 +0000 UTC m=+157.413699093" watchObservedRunningTime="2025-09-30 14:22:15.330468127 +0000 UTC m=+157.414068554" Sep 30 14:22:15 crc kubenswrapper[4799]: I0930 14:22:15.417728 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vxm25" Sep 30 14:22:16 crc kubenswrapper[4799]: I0930 14:22:16.283057 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:16 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:16 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:16 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:16 crc kubenswrapper[4799]: I0930 14:22:16.283126 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.035162 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.094321 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kubelet-dir\") pod \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\" (UID: \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\") " Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.094380 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kube-api-access\") pod \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\" (UID: \"ec017ab4-5210-40d6-a33a-ca7ace14bafe\") " Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.094394 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ec017ab4-5210-40d6-a33a-ca7ace14bafe" (UID: "ec017ab4-5210-40d6-a33a-ca7ace14bafe"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.094745 4799 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.108378 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ec017ab4-5210-40d6-a33a-ca7ace14bafe" (UID: "ec017ab4-5210-40d6-a33a-ca7ace14bafe"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.160486 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.196257 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access\") pod \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.198814 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kubelet-dir\") pod \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\" (UID: \"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0\") " Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.199753 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0" (UID: "bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.204910 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec017ab4-5210-40d6-a33a-ca7ace14bafe-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.204943 4799 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.207052 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0" (UID: "bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.284772 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:17 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:17 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:17 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.284891 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.306206 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.359417 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ec017ab4-5210-40d6-a33a-ca7ace14bafe","Type":"ContainerDied","Data":"c6cf7369eb5a7c4fa2cd1f3e1f2fc48793ee3594201c41accd464dc6b11a5127"} Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.359466 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6cf7369eb5a7c4fa2cd1f3e1f2fc48793ee3594201c41accd464dc6b11a5127" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.359560 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.387699 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0","Type":"ContainerDied","Data":"97887ffdbe31e0279232737c494a41510457cc30b75805e5b8fa4cfb520e2870"} Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.387786 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97887ffdbe31e0279232737c494a41510457cc30b75805e5b8fa4cfb520e2870" Sep 30 14:22:17 crc kubenswrapper[4799]: I0930 14:22:17.387890 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 30 14:22:18 crc kubenswrapper[4799]: I0930 14:22:18.279311 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:18 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:18 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:18 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:18 crc kubenswrapper[4799]: I0930 14:22:18.279724 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:18 crc kubenswrapper[4799]: I0930 14:22:18.975509 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:18 crc kubenswrapper[4799]: I0930 14:22:18.975585 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:18 crc kubenswrapper[4799]: I0930 14:22:18.980218 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:18 crc kubenswrapper[4799]: I0930 14:22:18.980501 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:19 crc kubenswrapper[4799]: I0930 14:22:19.286310 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:19 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:19 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:19 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:19 crc kubenswrapper[4799]: I0930 14:22:19.286377 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:19 crc kubenswrapper[4799]: I0930 14:22:19.439760 4799 generic.go:334] "Generic (PLEG): container finished" podID="7f147ef4-6fba-40b2-bdda-f19618d512eb" containerID="fff6ff50d0661449750a3403f5d699e084f7b63e133787d92d17e96c512110dd" exitCode=0 Sep 30 14:22:19 crc kubenswrapper[4799]: I0930 14:22:19.439809 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" event={"ID":"7f147ef4-6fba-40b2-bdda-f19618d512eb","Type":"ContainerDied","Data":"fff6ff50d0661449750a3403f5d699e084f7b63e133787d92d17e96c512110dd"} Sep 30 14:22:20 crc kubenswrapper[4799]: I0930 14:22:20.189176 4799 patch_prober.go:28] interesting pod/console-f9d7485db-svk57 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Sep 30 14:22:20 crc kubenswrapper[4799]: I0930 14:22:20.189548 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-svk57" podUID="03c9ee3f-722e-4db8-a8f5-dc359486e4f4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Sep 30 14:22:20 crc kubenswrapper[4799]: I0930 14:22:20.283936 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:20 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:20 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:20 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:20 crc kubenswrapper[4799]: I0930 14:22:20.283999 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:20 crc kubenswrapper[4799]: I0930 14:22:20.349317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:22:20 crc kubenswrapper[4799]: I0930 14:22:20.578054 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:22:20 crc kubenswrapper[4799]: I0930 14:22:20.584598 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15ddd33f-0b7f-4cb5-b09a-03c211c759c8-metrics-certs\") pod \"network-metrics-daemon-hghbr\" (UID: \"15ddd33f-0b7f-4cb5-b09a-03c211c759c8\") " pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:22:20 crc kubenswrapper[4799]: I0930 14:22:20.718772 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hghbr" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.207288 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.280075 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:21 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:21 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:21 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.280139 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.308278 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f147ef4-6fba-40b2-bdda-f19618d512eb-config-volume\") pod \"7f147ef4-6fba-40b2-bdda-f19618d512eb\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.308325 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qxxn\" (UniqueName: \"kubernetes.io/projected/7f147ef4-6fba-40b2-bdda-f19618d512eb-kube-api-access-9qxxn\") pod \"7f147ef4-6fba-40b2-bdda-f19618d512eb\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.308352 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f147ef4-6fba-40b2-bdda-f19618d512eb-secret-volume\") pod \"7f147ef4-6fba-40b2-bdda-f19618d512eb\" (UID: \"7f147ef4-6fba-40b2-bdda-f19618d512eb\") " Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.310313 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f147ef4-6fba-40b2-bdda-f19618d512eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "7f147ef4-6fba-40b2-bdda-f19618d512eb" (UID: "7f147ef4-6fba-40b2-bdda-f19618d512eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.322572 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f147ef4-6fba-40b2-bdda-f19618d512eb-kube-api-access-9qxxn" (OuterVolumeSpecName: "kube-api-access-9qxxn") pod "7f147ef4-6fba-40b2-bdda-f19618d512eb" (UID: "7f147ef4-6fba-40b2-bdda-f19618d512eb"). InnerVolumeSpecName "kube-api-access-9qxxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.322906 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f147ef4-6fba-40b2-bdda-f19618d512eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7f147ef4-6fba-40b2-bdda-f19618d512eb" (UID: "7f147ef4-6fba-40b2-bdda-f19618d512eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.410105 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f147ef4-6fba-40b2-bdda-f19618d512eb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.410138 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qxxn\" (UniqueName: \"kubernetes.io/projected/7f147ef4-6fba-40b2-bdda-f19618d512eb-kube-api-access-9qxxn\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.410149 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f147ef4-6fba-40b2-bdda-f19618d512eb-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.483375 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hghbr"] Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.602173 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-plkz9_8209103b-69e0-41be-9ce9-c87dd3e01a19/cluster-samples-operator/0.log" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.602569 4799 generic.go:334] "Generic (PLEG): container finished" podID="8209103b-69e0-41be-9ce9-c87dd3e01a19" containerID="cf6b484c78974acf164a204594123403bc01feef62825e33058f36e7b34966cf" exitCode=2 Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.602701 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" event={"ID":"8209103b-69e0-41be-9ce9-c87dd3e01a19","Type":"ContainerDied","Data":"cf6b484c78974acf164a204594123403bc01feef62825e33058f36e7b34966cf"} Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.603846 4799 scope.go:117] "RemoveContainer" containerID="cf6b484c78974acf164a204594123403bc01feef62825e33058f36e7b34966cf" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.622453 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" event={"ID":"7f147ef4-6fba-40b2-bdda-f19618d512eb","Type":"ContainerDied","Data":"080b1461b14fb7b7e7794967baafbf62538826858cabdadf8a6a8811eb4272a3"} Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.622490 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="080b1461b14fb7b7e7794967baafbf62538826858cabdadf8a6a8811eb4272a3" Sep 30 14:22:21 crc kubenswrapper[4799]: I0930 14:22:21.622549 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t" Sep 30 14:22:22 crc kubenswrapper[4799]: I0930 14:22:22.284800 4799 patch_prober.go:28] interesting pod/router-default-5444994796-glzch container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 30 14:22:22 crc kubenswrapper[4799]: [-]has-synced failed: reason withheld Sep 30 14:22:22 crc kubenswrapper[4799]: [+]process-running ok Sep 30 14:22:22 crc kubenswrapper[4799]: healthz check failed Sep 30 14:22:22 crc kubenswrapper[4799]: I0930 14:22:22.284877 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-glzch" podUID="6e587e9f-8d37-46ef-ab7e-23e5f772078b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 30 14:22:22 crc kubenswrapper[4799]: I0930 14:22:22.638012 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hghbr" event={"ID":"15ddd33f-0b7f-4cb5-b09a-03c211c759c8","Type":"ContainerStarted","Data":"a09a817b8b6323eb61d2f1aeb7f96b5c588c667a15f89e0b5a6a7fe1ef6c3171"} Sep 30 14:22:22 crc kubenswrapper[4799]: I0930 14:22:22.650891 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-plkz9_8209103b-69e0-41be-9ce9-c87dd3e01a19/cluster-samples-operator/0.log" Sep 30 14:22:22 crc kubenswrapper[4799]: I0930 14:22:22.650945 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-plkz9" event={"ID":"8209103b-69e0-41be-9ce9-c87dd3e01a19","Type":"ContainerStarted","Data":"f46dbbc357704c661d898d534f1496a06264b3233a4816fee2f3b497c27ce7d8"} Sep 30 14:22:23 crc kubenswrapper[4799]: I0930 14:22:23.286834 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:23 crc kubenswrapper[4799]: I0930 14:22:23.291714 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-glzch" Sep 30 14:22:23 crc kubenswrapper[4799]: I0930 14:22:23.695393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hghbr" event={"ID":"15ddd33f-0b7f-4cb5-b09a-03c211c759c8","Type":"ContainerStarted","Data":"e9875c911f981656439418c0027eb9e427f182bf0bf5171064cfcc9a4c5464d1"} Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.977457 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.977774 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.977815 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.978305 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"0fb9f012503196c07acd46c85420f5f64dbd0054be968eaac1851ba85a316966"} pod="openshift-console/downloads-7954f5f757-7f9sl" containerMessage="Container download-server failed liveness probe, will be restarted" Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.978387 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" containerID="cri-o://0fb9f012503196c07acd46c85420f5f64dbd0054be968eaac1851ba85a316966" gracePeriod=2 Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.977470 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.978517 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.978930 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:28 crc kubenswrapper[4799]: I0930 14:22:28.978946 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:29 crc kubenswrapper[4799]: I0930 14:22:29.648818 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:22:29 crc kubenswrapper[4799]: I0930 14:22:29.648882 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:22:29 crc kubenswrapper[4799]: I0930 14:22:29.839930 4799 generic.go:334] "Generic (PLEG): container finished" podID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerID="0fb9f012503196c07acd46c85420f5f64dbd0054be968eaac1851ba85a316966" exitCode=0 Sep 30 14:22:29 crc kubenswrapper[4799]: I0930 14:22:29.839979 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7f9sl" event={"ID":"e2e2061e-3d24-41dd-b8ee-6b50e8313647","Type":"ContainerDied","Data":"0fb9f012503196c07acd46c85420f5f64dbd0054be968eaac1851ba85a316966"} Sep 30 14:22:30 crc kubenswrapper[4799]: I0930 14:22:30.196614 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:22:30 crc kubenswrapper[4799]: I0930 14:22:30.200768 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:22:34 crc kubenswrapper[4799]: I0930 14:22:34.016534 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:22:38 crc kubenswrapper[4799]: I0930 14:22:38.975560 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:38 crc kubenswrapper[4799]: I0930 14:22:38.976065 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:40 crc kubenswrapper[4799]: I0930 14:22:40.320224 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-77w76" Sep 30 14:22:47 crc kubenswrapper[4799]: I0930 14:22:47.828866 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 30 14:22:48 crc kubenswrapper[4799]: I0930 14:22:48.975223 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:48 crc kubenswrapper[4799]: I0930 14:22:48.975273 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:52 crc kubenswrapper[4799]: E0930 14:22:52.661630 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 30 14:22:52 crc kubenswrapper[4799]: E0930 14:22:52.662314 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4qfl7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pdhnm_openshift-marketplace(b9eab5bb-9ccf-4184-a82e-25b41844ea1b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 14:22:52 crc kubenswrapper[4799]: E0930 14:22:52.663467 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pdhnm" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" Sep 30 14:22:53 crc kubenswrapper[4799]: E0930 14:22:53.663059 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pdhnm" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" Sep 30 14:22:53 crc kubenswrapper[4799]: E0930 14:22:53.724438 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 30 14:22:53 crc kubenswrapper[4799]: E0930 14:22:53.724590 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lgh5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pzj2v_openshift-marketplace(fe8c96af-648f-407f-a61b-d09601743327): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 14:22:53 crc kubenswrapper[4799]: E0930 14:22:53.725991 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pzj2v" podUID="fe8c96af-648f-407f-a61b-d09601743327" Sep 30 14:22:54 crc kubenswrapper[4799]: E0930 14:22:54.937615 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-pzj2v" podUID="fe8c96af-648f-407f-a61b-d09601743327" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.015097 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.015286 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-blpcv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zrqpk_openshift-marketplace(70b0c843-7e0f-4cca-8443-9306f0c7ad76): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.016719 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zrqpk" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.037593 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.037763 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5f8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-76bw2_openshift-marketplace(4eb004e8-ef9d-404c-9321-2c566b86e393): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.039112 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-76bw2" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.090315 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.090467 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d8flc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-nh8v4_openshift-marketplace(5ef93211-55a5-43d0-9bda-a1e41c44fb35): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.091802 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-nh8v4" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.199960 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.200540 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h7c6f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9vmjm_openshift-marketplace(5b9dfdd6-a206-4a32-8e66-4402828d4d6c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 14:22:55 crc kubenswrapper[4799]: E0930 14:22:55.201865 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9vmjm" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.005830 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hghbr" event={"ID":"15ddd33f-0b7f-4cb5-b09a-03c211c759c8","Type":"ContainerStarted","Data":"96049aa825233e17f7cd18b3bb29a98f7fab811dd3603de29a06b70ddb9c1904"} Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.008986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntf96" event={"ID":"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f","Type":"ContainerStarted","Data":"6e49d00cb0f35ee40f271b85e9d53c4ab7309818d02d7f31ab65426006d6b626"} Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.014356 4799 generic.go:334] "Generic (PLEG): container finished" podID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerID="765af5ef5a37b37e375b12aaa903bb8e88a708bc80e9d704d00839c35c9de38f" exitCode=0 Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.014402 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2q" event={"ID":"674513e3-aa0c-4814-884c-d0cb1468a3c9","Type":"ContainerDied","Data":"765af5ef5a37b37e375b12aaa903bb8e88a708bc80e9d704d00839c35c9de38f"} Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.019201 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7f9sl" event={"ID":"e2e2061e-3d24-41dd-b8ee-6b50e8313647","Type":"ContainerStarted","Data":"232656079d3a95bd609edb47ddef389dbec634a2ccd68184438dbb7ca0439705"} Sep 30 14:22:56 crc kubenswrapper[4799]: E0930 14:22:56.019349 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-nh8v4" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.019967 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.022015 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.022100 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:56 crc kubenswrapper[4799]: E0930 14:22:56.022351 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9vmjm" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" Sep 30 14:22:56 crc kubenswrapper[4799]: E0930 14:22:56.022429 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zrqpk" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" Sep 30 14:22:56 crc kubenswrapper[4799]: E0930 14:22:56.022742 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-76bw2" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" Sep 30 14:22:56 crc kubenswrapper[4799]: I0930 14:22:56.034264 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hghbr" podStartSLOduration=178.034240137 podStartE2EDuration="2m58.034240137s" podCreationTimestamp="2025-09-30 14:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:22:56.033700229 +0000 UTC m=+198.117300666" watchObservedRunningTime="2025-09-30 14:22:56.034240137 +0000 UTC m=+198.117840554" Sep 30 14:22:57 crc kubenswrapper[4799]: I0930 14:22:57.029721 4799 generic.go:334] "Generic (PLEG): container finished" podID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerID="6e49d00cb0f35ee40f271b85e9d53c4ab7309818d02d7f31ab65426006d6b626" exitCode=0 Sep 30 14:22:57 crc kubenswrapper[4799]: I0930 14:22:57.029804 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntf96" event={"ID":"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f","Type":"ContainerDied","Data":"6e49d00cb0f35ee40f271b85e9d53c4ab7309818d02d7f31ab65426006d6b626"} Sep 30 14:22:57 crc kubenswrapper[4799]: I0930 14:22:57.032843 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:57 crc kubenswrapper[4799]: I0930 14:22:57.032880 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:58 crc kubenswrapper[4799]: I0930 14:22:58.038137 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:58 crc kubenswrapper[4799]: I0930 14:22:58.038206 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:58 crc kubenswrapper[4799]: I0930 14:22:58.974603 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:58 crc kubenswrapper[4799]: I0930 14:22:58.974684 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:58 crc kubenswrapper[4799]: I0930 14:22:58.974930 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:22:58 crc kubenswrapper[4799]: I0930 14:22:58.974959 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:22:59 crc kubenswrapper[4799]: I0930 14:22:59.648584 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:22:59 crc kubenswrapper[4799]: I0930 14:22:59.648637 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:23:02 crc kubenswrapper[4799]: I0930 14:23:02.060723 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2q" event={"ID":"674513e3-aa0c-4814-884c-d0cb1468a3c9","Type":"ContainerStarted","Data":"6ee699983927862c3a98c1a550fc8f69384bd31cfe40f5202e623df00660f923"} Sep 30 14:23:02 crc kubenswrapper[4799]: I0930 14:23:02.081502 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nzj2q" podStartSLOduration=7.369201311 podStartE2EDuration="54.081483529s" podCreationTimestamp="2025-09-30 14:22:08 +0000 UTC" firstStartedPulling="2025-09-30 14:22:14.075451687 +0000 UTC m=+156.159052114" lastFinishedPulling="2025-09-30 14:23:00.787733915 +0000 UTC m=+202.871334332" observedRunningTime="2025-09-30 14:23:02.076909365 +0000 UTC m=+204.160509812" watchObservedRunningTime="2025-09-30 14:23:02.081483529 +0000 UTC m=+204.165083956" Sep 30 14:23:04 crc kubenswrapper[4799]: I0930 14:23:04.071878 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntf96" event={"ID":"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f","Type":"ContainerStarted","Data":"bce9dcbc749673a2934143c43187441ae7361083f1ae3238810d3e41cb56c974"} Sep 30 14:23:04 crc kubenswrapper[4799]: I0930 14:23:04.093379 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ntf96" podStartSLOduration=4.935940989 podStartE2EDuration="53.093361975s" podCreationTimestamp="2025-09-30 14:22:11 +0000 UTC" firstStartedPulling="2025-09-30 14:22:15.234815885 +0000 UTC m=+157.318416312" lastFinishedPulling="2025-09-30 14:23:03.392236871 +0000 UTC m=+205.475837298" observedRunningTime="2025-09-30 14:23:04.091003746 +0000 UTC m=+206.174604203" watchObservedRunningTime="2025-09-30 14:23:04.093361975 +0000 UTC m=+206.176962402" Sep 30 14:23:08 crc kubenswrapper[4799]: I0930 14:23:08.974917 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:23:08 crc kubenswrapper[4799]: I0930 14:23:08.975536 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:23:08 crc kubenswrapper[4799]: I0930 14:23:08.975011 4799 patch_prober.go:28] interesting pod/downloads-7954f5f757-7f9sl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Sep 30 14:23:08 crc kubenswrapper[4799]: I0930 14:23:08.975877 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7f9sl" podUID="e2e2061e-3d24-41dd-b8ee-6b50e8313647" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Sep 30 14:23:09 crc kubenswrapper[4799]: I0930 14:23:09.410673 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:23:09 crc kubenswrapper[4799]: I0930 14:23:09.410726 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:23:10 crc kubenswrapper[4799]: I0930 14:23:10.083483 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:23:10 crc kubenswrapper[4799]: I0930 14:23:10.154973 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:23:11 crc kubenswrapper[4799]: I0930 14:23:11.112416 4799 generic.go:334] "Generic (PLEG): container finished" podID="fe8c96af-648f-407f-a61b-d09601743327" containerID="5e5d59325a73fdfa68a2e35c7f23840faf3c6611e663d3124b07e5cc491b99b6" exitCode=0 Sep 30 14:23:11 crc kubenswrapper[4799]: I0930 14:23:11.112590 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pzj2v" event={"ID":"fe8c96af-648f-407f-a61b-d09601743327","Type":"ContainerDied","Data":"5e5d59325a73fdfa68a2e35c7f23840faf3c6611e663d3124b07e5cc491b99b6"} Sep 30 14:23:11 crc kubenswrapper[4799]: I0930 14:23:11.122826 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdhnm" event={"ID":"b9eab5bb-9ccf-4184-a82e-25b41844ea1b","Type":"ContainerStarted","Data":"e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55"} Sep 30 14:23:12 crc kubenswrapper[4799]: I0930 14:23:12.131210 4799 generic.go:334] "Generic (PLEG): container finished" podID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerID="e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55" exitCode=0 Sep 30 14:23:12 crc kubenswrapper[4799]: I0930 14:23:12.131433 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdhnm" event={"ID":"b9eab5bb-9ccf-4184-a82e-25b41844ea1b","Type":"ContainerDied","Data":"e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55"} Sep 30 14:23:12 crc kubenswrapper[4799]: I0930 14:23:12.134639 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76bw2" event={"ID":"4eb004e8-ef9d-404c-9321-2c566b86e393","Type":"ContainerStarted","Data":"c7e5ac575f406dbc52a27b109749f519eb14c02a6f77c06db747b0ff3c8024aa"} Sep 30 14:23:12 crc kubenswrapper[4799]: I0930 14:23:12.136940 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh8v4" event={"ID":"5ef93211-55a5-43d0-9bda-a1e41c44fb35","Type":"ContainerStarted","Data":"cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c"} Sep 30 14:23:12 crc kubenswrapper[4799]: I0930 14:23:12.140178 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vmjm" event={"ID":"5b9dfdd6-a206-4a32-8e66-4402828d4d6c","Type":"ContainerStarted","Data":"72bc9602034e089cc8b852a788f2a499f2f8b29b1ff6ddecf66cce168dbaef6a"} Sep 30 14:23:12 crc kubenswrapper[4799]: I0930 14:23:12.629174 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:23:12 crc kubenswrapper[4799]: I0930 14:23:12.629212 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:23:12 crc kubenswrapper[4799]: I0930 14:23:12.678506 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:23:13 crc kubenswrapper[4799]: I0930 14:23:13.148954 4799 generic.go:334] "Generic (PLEG): container finished" podID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerID="72bc9602034e089cc8b852a788f2a499f2f8b29b1ff6ddecf66cce168dbaef6a" exitCode=0 Sep 30 14:23:13 crc kubenswrapper[4799]: I0930 14:23:13.149403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vmjm" event={"ID":"5b9dfdd6-a206-4a32-8e66-4402828d4d6c","Type":"ContainerDied","Data":"72bc9602034e089cc8b852a788f2a499f2f8b29b1ff6ddecf66cce168dbaef6a"} Sep 30 14:23:13 crc kubenswrapper[4799]: I0930 14:23:13.157558 4799 generic.go:334] "Generic (PLEG): container finished" podID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerID="cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c" exitCode=0 Sep 30 14:23:13 crc kubenswrapper[4799]: I0930 14:23:13.157629 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh8v4" event={"ID":"5ef93211-55a5-43d0-9bda-a1e41c44fb35","Type":"ContainerDied","Data":"cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c"} Sep 30 14:23:13 crc kubenswrapper[4799]: I0930 14:23:13.185045 4799 generic.go:334] "Generic (PLEG): container finished" podID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerID="c7e5ac575f406dbc52a27b109749f519eb14c02a6f77c06db747b0ff3c8024aa" exitCode=0 Sep 30 14:23:13 crc kubenswrapper[4799]: I0930 14:23:13.185142 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76bw2" event={"ID":"4eb004e8-ef9d-404c-9321-2c566b86e393","Type":"ContainerDied","Data":"c7e5ac575f406dbc52a27b109749f519eb14c02a6f77c06db747b0ff3c8024aa"} Sep 30 14:23:13 crc kubenswrapper[4799]: I0930 14:23:13.244144 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:23:17 crc kubenswrapper[4799]: I0930 14:23:17.210684 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pzj2v" event={"ID":"fe8c96af-648f-407f-a61b-d09601743327","Type":"ContainerStarted","Data":"340fb776cf7a8f5797e0ff4b269d1eafe1708022ea5de25b9b1c85e6f58062a0"} Sep 30 14:23:17 crc kubenswrapper[4799]: I0930 14:23:17.232099 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pzj2v" podStartSLOduration=6.264094244 podStartE2EDuration="1m7.232082562s" podCreationTimestamp="2025-09-30 14:22:10 +0000 UTC" firstStartedPulling="2025-09-30 14:22:14.191137011 +0000 UTC m=+156.274737438" lastFinishedPulling="2025-09-30 14:23:15.159125319 +0000 UTC m=+217.242725756" observedRunningTime="2025-09-30 14:23:17.231010506 +0000 UTC m=+219.314610963" watchObservedRunningTime="2025-09-30 14:23:17.232082562 +0000 UTC m=+219.315682989" Sep 30 14:23:18 crc kubenswrapper[4799]: I0930 14:23:18.217837 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdhnm" event={"ID":"b9eab5bb-9ccf-4184-a82e-25b41844ea1b","Type":"ContainerStarted","Data":"6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65"} Sep 30 14:23:18 crc kubenswrapper[4799]: I0930 14:23:18.239052 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pdhnm" podStartSLOduration=5.367165726 podStartE2EDuration="1m7.239030449s" podCreationTimestamp="2025-09-30 14:22:11 +0000 UTC" firstStartedPulling="2025-09-30 14:22:15.246283192 +0000 UTC m=+157.329883619" lastFinishedPulling="2025-09-30 14:23:17.118147925 +0000 UTC m=+219.201748342" observedRunningTime="2025-09-30 14:23:18.234019081 +0000 UTC m=+220.317619518" watchObservedRunningTime="2025-09-30 14:23:18.239030449 +0000 UTC m=+220.322630876" Sep 30 14:23:18 crc kubenswrapper[4799]: I0930 14:23:18.982780 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-7f9sl" Sep 30 14:23:21 crc kubenswrapper[4799]: I0930 14:23:21.324447 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:23:21 crc kubenswrapper[4799]: I0930 14:23:21.325031 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:23:21 crc kubenswrapper[4799]: I0930 14:23:21.377508 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.244016 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh8v4" event={"ID":"5ef93211-55a5-43d0-9bda-a1e41c44fb35","Type":"ContainerStarted","Data":"a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4"} Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.246632 4799 generic.go:334] "Generic (PLEG): container finished" podID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerID="0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0" exitCode=0 Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.246688 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrqpk" event={"ID":"70b0c843-7e0f-4cca-8443-9306f0c7ad76","Type":"ContainerDied","Data":"0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0"} Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.255843 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vmjm" event={"ID":"5b9dfdd6-a206-4a32-8e66-4402828d4d6c","Type":"ContainerStarted","Data":"7edd06435a6f3b27882013d5b3fe1b4c859055bc0fd0ea83d39901ac235b3a04"} Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.261605 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76bw2" event={"ID":"4eb004e8-ef9d-404c-9321-2c566b86e393","Type":"ContainerStarted","Data":"1c94223e7ae062e9c81bc1e246eefe005753ef17edc4fde6d1ea9ca233b9f2be"} Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.277174 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nh8v4" podStartSLOduration=3.8293204960000002 podStartE2EDuration="1m11.277153955s" podCreationTimestamp="2025-09-30 14:22:11 +0000 UTC" firstStartedPulling="2025-09-30 14:22:14.090059854 +0000 UTC m=+156.173660281" lastFinishedPulling="2025-09-30 14:23:21.537893313 +0000 UTC m=+223.621493740" observedRunningTime="2025-09-30 14:23:22.272762958 +0000 UTC m=+224.356363385" watchObservedRunningTime="2025-09-30 14:23:22.277153955 +0000 UTC m=+224.360754382" Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.302485 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-76bw2" podStartSLOduration=5.8252782199999995 podStartE2EDuration="1m14.302471823s" podCreationTimestamp="2025-09-30 14:22:08 +0000 UTC" firstStartedPulling="2025-09-30 14:22:13.062662886 +0000 UTC m=+155.146263313" lastFinishedPulling="2025-09-30 14:23:21.539856479 +0000 UTC m=+223.623456916" observedRunningTime="2025-09-30 14:23:22.302154122 +0000 UTC m=+224.385754549" watchObservedRunningTime="2025-09-30 14:23:22.302471823 +0000 UTC m=+224.386072250" Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.328713 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9vmjm" podStartSLOduration=6.072670736 podStartE2EDuration="1m13.328689721s" podCreationTimestamp="2025-09-30 14:22:09 +0000 UTC" firstStartedPulling="2025-09-30 14:22:14.19983266 +0000 UTC m=+156.283433087" lastFinishedPulling="2025-09-30 14:23:21.455851645 +0000 UTC m=+223.539452072" observedRunningTime="2025-09-30 14:23:22.325309978 +0000 UTC m=+224.408910425" watchObservedRunningTime="2025-09-30 14:23:22.328689721 +0000 UTC m=+224.412290158" Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.380010 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.487495 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:23:22 crc kubenswrapper[4799]: I0930 14:23:22.487541 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:23:23 crc kubenswrapper[4799]: I0930 14:23:23.542395 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pdhnm" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="registry-server" probeResult="failure" output=< Sep 30 14:23:23 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 14:23:23 crc kubenswrapper[4799]: > Sep 30 14:23:24 crc kubenswrapper[4799]: I0930 14:23:24.274992 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrqpk" event={"ID":"70b0c843-7e0f-4cca-8443-9306f0c7ad76","Type":"ContainerStarted","Data":"52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02"} Sep 30 14:23:24 crc kubenswrapper[4799]: I0930 14:23:24.293604 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zrqpk" podStartSLOduration=7.029669642 podStartE2EDuration="1m16.293583834s" podCreationTimestamp="2025-09-30 14:22:08 +0000 UTC" firstStartedPulling="2025-09-30 14:22:14.10286279 +0000 UTC m=+156.186463217" lastFinishedPulling="2025-09-30 14:23:23.366776982 +0000 UTC m=+225.450377409" observedRunningTime="2025-09-30 14:23:24.291393331 +0000 UTC m=+226.374993788" watchObservedRunningTime="2025-09-30 14:23:24.293583834 +0000 UTC m=+226.377184271" Sep 30 14:23:26 crc kubenswrapper[4799]: I0930 14:23:26.882995 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfxv2"] Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.218256 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.218670 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.275381 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.280905 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.280946 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.326843 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.350464 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.387096 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.586995 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.587336 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.630419 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.649221 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.649287 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.649352 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.650007 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:23:29 crc kubenswrapper[4799]: I0930 14:23:29.650082 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa" gracePeriod=600 Sep 30 14:23:30 crc kubenswrapper[4799]: I0930 14:23:30.309160 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa" exitCode=0 Sep 30 14:23:30 crc kubenswrapper[4799]: I0930 14:23:30.309202 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa"} Sep 30 14:23:30 crc kubenswrapper[4799]: I0930 14:23:30.310495 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"104fd6ad26a2612596b50bf918e18ccbe2581dacca6d978ddfbd4cf3ad4e5322"} Sep 30 14:23:30 crc kubenswrapper[4799]: I0930 14:23:30.360447 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:23:30 crc kubenswrapper[4799]: I0930 14:23:30.541249 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrqpk"] Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.318952 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zrqpk" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerName="registry-server" containerID="cri-o://52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02" gracePeriod=2 Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.690815 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.730083 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-utilities\") pod \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.730629 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blpcv\" (UniqueName: \"kubernetes.io/projected/70b0c843-7e0f-4cca-8443-9306f0c7ad76-kube-api-access-blpcv\") pod \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.730682 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-catalog-content\") pod \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\" (UID: \"70b0c843-7e0f-4cca-8443-9306f0c7ad76\") " Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.733499 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-utilities" (OuterVolumeSpecName: "utilities") pod "70b0c843-7e0f-4cca-8443-9306f0c7ad76" (UID: "70b0c843-7e0f-4cca-8443-9306f0c7ad76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.742810 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70b0c843-7e0f-4cca-8443-9306f0c7ad76-kube-api-access-blpcv" (OuterVolumeSpecName: "kube-api-access-blpcv") pod "70b0c843-7e0f-4cca-8443-9306f0c7ad76" (UID: "70b0c843-7e0f-4cca-8443-9306f0c7ad76"). InnerVolumeSpecName "kube-api-access-blpcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.793028 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70b0c843-7e0f-4cca-8443-9306f0c7ad76" (UID: "70b0c843-7e0f-4cca-8443-9306f0c7ad76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.832105 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.832150 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blpcv\" (UniqueName: \"kubernetes.io/projected/70b0c843-7e0f-4cca-8443-9306f0c7ad76-kube-api-access-blpcv\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:31 crc kubenswrapper[4799]: I0930 14:23:31.832161 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70b0c843-7e0f-4cca-8443-9306f0c7ad76-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.103268 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.103337 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.164808 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.341872 4799 generic.go:334] "Generic (PLEG): container finished" podID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerID="52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02" exitCode=0 Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.342087 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrqpk" event={"ID":"70b0c843-7e0f-4cca-8443-9306f0c7ad76","Type":"ContainerDied","Data":"52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02"} Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.342215 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrqpk" event={"ID":"70b0c843-7e0f-4cca-8443-9306f0c7ad76","Type":"ContainerDied","Data":"cb1e8b5dc9538e9d403d0cbc5c33d419834d80a3a7378e6b4dd4ada7f56a7ea3"} Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.342227 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrqpk" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.342246 4799 scope.go:117] "RemoveContainer" containerID="52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.381816 4799 scope.go:117] "RemoveContainer" containerID="0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.401282 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrqpk"] Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.410421 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zrqpk"] Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.411017 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.423022 4799 scope.go:117] "RemoveContainer" containerID="962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.448227 4799 scope.go:117] "RemoveContainer" containerID="52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02" Sep 30 14:23:32 crc kubenswrapper[4799]: E0930 14:23:32.448875 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02\": container with ID starting with 52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02 not found: ID does not exist" containerID="52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.448948 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02"} err="failed to get container status \"52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02\": rpc error: code = NotFound desc = could not find container \"52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02\": container with ID starting with 52acd4fe8cffa8bb3837fe8ef9afd400aa6b821bfd28cb1601730a44a51a1d02 not found: ID does not exist" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.448996 4799 scope.go:117] "RemoveContainer" containerID="0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0" Sep 30 14:23:32 crc kubenswrapper[4799]: E0930 14:23:32.449474 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0\": container with ID starting with 0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0 not found: ID does not exist" containerID="0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.449507 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0"} err="failed to get container status \"0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0\": rpc error: code = NotFound desc = could not find container \"0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0\": container with ID starting with 0dd49c56d32281b6c72842aca9c4e572c4986049e622d8480b12c48265e5e2e0 not found: ID does not exist" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.449526 4799 scope.go:117] "RemoveContainer" containerID="962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f" Sep 30 14:23:32 crc kubenswrapper[4799]: E0930 14:23:32.450198 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f\": container with ID starting with 962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f not found: ID does not exist" containerID="962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.450224 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f"} err="failed to get container status \"962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f\": rpc error: code = NotFound desc = could not find container \"962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f\": container with ID starting with 962e5a48cdbd86507937b02a9c8c67d395ee49163a680488eb3168e3437f7f1f not found: ID does not exist" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.510997 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" path="/var/lib/kubelet/pods/70b0c843-7e0f-4cca-8443-9306f0c7ad76/volumes" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.533723 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.580017 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.944497 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9vmjm"] Sep 30 14:23:32 crc kubenswrapper[4799]: I0930 14:23:32.945038 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9vmjm" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerName="registry-server" containerID="cri-o://7edd06435a6f3b27882013d5b3fe1b4c859055bc0fd0ea83d39901ac235b3a04" gracePeriod=2 Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.139924 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh8v4"] Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.352444 4799 generic.go:334] "Generic (PLEG): container finished" podID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerID="7edd06435a6f3b27882013d5b3fe1b4c859055bc0fd0ea83d39901ac235b3a04" exitCode=0 Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.352555 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vmjm" event={"ID":"5b9dfdd6-a206-4a32-8e66-4402828d4d6c","Type":"ContainerDied","Data":"7edd06435a6f3b27882013d5b3fe1b4c859055bc0fd0ea83d39901ac235b3a04"} Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.352592 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vmjm" event={"ID":"5b9dfdd6-a206-4a32-8e66-4402828d4d6c","Type":"ContainerDied","Data":"882e9d4819e4f5f61dd09803f22b1cf1a804c1a3e4cb1535f53337f9e68e662f"} Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.352604 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="882e9d4819e4f5f61dd09803f22b1cf1a804c1a3e4cb1535f53337f9e68e662f" Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.372953 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.452420 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-utilities\") pod \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.452543 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-catalog-content\") pod \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.452638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7c6f\" (UniqueName: \"kubernetes.io/projected/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-kube-api-access-h7c6f\") pod \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\" (UID: \"5b9dfdd6-a206-4a32-8e66-4402828d4d6c\") " Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.453169 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-utilities" (OuterVolumeSpecName: "utilities") pod "5b9dfdd6-a206-4a32-8e66-4402828d4d6c" (UID: "5b9dfdd6-a206-4a32-8e66-4402828d4d6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.458828 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-kube-api-access-h7c6f" (OuterVolumeSpecName: "kube-api-access-h7c6f") pod "5b9dfdd6-a206-4a32-8e66-4402828d4d6c" (UID: "5b9dfdd6-a206-4a32-8e66-4402828d4d6c"). InnerVolumeSpecName "kube-api-access-h7c6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.514746 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b9dfdd6-a206-4a32-8e66-4402828d4d6c" (UID: "5b9dfdd6-a206-4a32-8e66-4402828d4d6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.554562 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.554823 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7c6f\" (UniqueName: \"kubernetes.io/projected/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-kube-api-access-h7c6f\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:33 crc kubenswrapper[4799]: I0930 14:23:33.554925 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b9dfdd6-a206-4a32-8e66-4402828d4d6c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.357667 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vmjm" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.358600 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nh8v4" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerName="registry-server" containerID="cri-o://a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4" gracePeriod=2 Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.390811 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9vmjm"] Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.393936 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9vmjm"] Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.510375 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" path="/var/lib/kubelet/pods/5b9dfdd6-a206-4a32-8e66-4402828d4d6c/volumes" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.751940 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.874554 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-catalog-content\") pod \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.874617 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-utilities\") pod \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.874735 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8flc\" (UniqueName: \"kubernetes.io/projected/5ef93211-55a5-43d0-9bda-a1e41c44fb35-kube-api-access-d8flc\") pod \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\" (UID: \"5ef93211-55a5-43d0-9bda-a1e41c44fb35\") " Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.876011 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-utilities" (OuterVolumeSpecName: "utilities") pod "5ef93211-55a5-43d0-9bda-a1e41c44fb35" (UID: "5ef93211-55a5-43d0-9bda-a1e41c44fb35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.879819 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef93211-55a5-43d0-9bda-a1e41c44fb35-kube-api-access-d8flc" (OuterVolumeSpecName: "kube-api-access-d8flc") pod "5ef93211-55a5-43d0-9bda-a1e41c44fb35" (UID: "5ef93211-55a5-43d0-9bda-a1e41c44fb35"). InnerVolumeSpecName "kube-api-access-d8flc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.888670 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ef93211-55a5-43d0-9bda-a1e41c44fb35" (UID: "5ef93211-55a5-43d0-9bda-a1e41c44fb35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.976061 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.976093 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef93211-55a5-43d0-9bda-a1e41c44fb35-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:34 crc kubenswrapper[4799]: I0930 14:23:34.976104 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8flc\" (UniqueName: \"kubernetes.io/projected/5ef93211-55a5-43d0-9bda-a1e41c44fb35-kube-api-access-d8flc\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.368806 4799 generic.go:334] "Generic (PLEG): container finished" podID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerID="a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4" exitCode=0 Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.368897 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nh8v4" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.368933 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh8v4" event={"ID":"5ef93211-55a5-43d0-9bda-a1e41c44fb35","Type":"ContainerDied","Data":"a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4"} Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.369489 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nh8v4" event={"ID":"5ef93211-55a5-43d0-9bda-a1e41c44fb35","Type":"ContainerDied","Data":"8520605c506fde2c7c70ad2634c72024cd3ac1cce9402efad2b06147f5855f4b"} Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.369514 4799 scope.go:117] "RemoveContainer" containerID="a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.397500 4799 scope.go:117] "RemoveContainer" containerID="cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.415255 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh8v4"] Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.422875 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nh8v4"] Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.430138 4799 scope.go:117] "RemoveContainer" containerID="c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.452862 4799 scope.go:117] "RemoveContainer" containerID="a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4" Sep 30 14:23:35 crc kubenswrapper[4799]: E0930 14:23:35.453419 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4\": container with ID starting with a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4 not found: ID does not exist" containerID="a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.453495 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4"} err="failed to get container status \"a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4\": rpc error: code = NotFound desc = could not find container \"a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4\": container with ID starting with a4aa38b92c24775474b817514a3ebbb6da47a6f2f599c1d8ea05c449a334a7f4 not found: ID does not exist" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.453535 4799 scope.go:117] "RemoveContainer" containerID="cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c" Sep 30 14:23:35 crc kubenswrapper[4799]: E0930 14:23:35.454333 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c\": container with ID starting with cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c not found: ID does not exist" containerID="cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.454369 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c"} err="failed to get container status \"cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c\": rpc error: code = NotFound desc = could not find container \"cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c\": container with ID starting with cffb0f027012dc7deea13c26cab0683a4d338475f19639db633d04aba140a21c not found: ID does not exist" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.454390 4799 scope.go:117] "RemoveContainer" containerID="c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7" Sep 30 14:23:35 crc kubenswrapper[4799]: E0930 14:23:35.454955 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7\": container with ID starting with c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7 not found: ID does not exist" containerID="c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.455014 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7"} err="failed to get container status \"c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7\": rpc error: code = NotFound desc = could not find container \"c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7\": container with ID starting with c50346feaa2f5f0097fbdf6fb8d82f392a3d2c6d9be53557f22271c030ac36c7 not found: ID does not exist" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.540637 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pdhnm"] Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.540904 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pdhnm" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="registry-server" containerID="cri-o://6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65" gracePeriod=2 Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.901533 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.989468 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qfl7\" (UniqueName: \"kubernetes.io/projected/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-kube-api-access-4qfl7\") pod \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.989554 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-utilities\") pod \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.989700 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-catalog-content\") pod \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\" (UID: \"b9eab5bb-9ccf-4184-a82e-25b41844ea1b\") " Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.990398 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-utilities" (OuterVolumeSpecName: "utilities") pod "b9eab5bb-9ccf-4184-a82e-25b41844ea1b" (UID: "b9eab5bb-9ccf-4184-a82e-25b41844ea1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:23:35 crc kubenswrapper[4799]: I0930 14:23:35.998811 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-kube-api-access-4qfl7" (OuterVolumeSpecName: "kube-api-access-4qfl7") pod "b9eab5bb-9ccf-4184-a82e-25b41844ea1b" (UID: "b9eab5bb-9ccf-4184-a82e-25b41844ea1b"). InnerVolumeSpecName "kube-api-access-4qfl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.074529 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9eab5bb-9ccf-4184-a82e-25b41844ea1b" (UID: "b9eab5bb-9ccf-4184-a82e-25b41844ea1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.091505 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qfl7\" (UniqueName: \"kubernetes.io/projected/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-kube-api-access-4qfl7\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.091536 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.091545 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9eab5bb-9ccf-4184-a82e-25b41844ea1b-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.376522 4799 generic.go:334] "Generic (PLEG): container finished" podID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerID="6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65" exitCode=0 Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.376581 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdhnm" event={"ID":"b9eab5bb-9ccf-4184-a82e-25b41844ea1b","Type":"ContainerDied","Data":"6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65"} Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.376589 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pdhnm" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.376666 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pdhnm" event={"ID":"b9eab5bb-9ccf-4184-a82e-25b41844ea1b","Type":"ContainerDied","Data":"3063ea6c02d5bce2c8bdf7498786ca50d53c797a66645c79d418fbdb90822ea2"} Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.376674 4799 scope.go:117] "RemoveContainer" containerID="6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.395129 4799 scope.go:117] "RemoveContainer" containerID="e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.406792 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pdhnm"] Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.411015 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pdhnm"] Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.418558 4799 scope.go:117] "RemoveContainer" containerID="4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.432559 4799 scope.go:117] "RemoveContainer" containerID="6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65" Sep 30 14:23:36 crc kubenswrapper[4799]: E0930 14:23:36.433020 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65\": container with ID starting with 6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65 not found: ID does not exist" containerID="6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.433068 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65"} err="failed to get container status \"6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65\": rpc error: code = NotFound desc = could not find container \"6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65\": container with ID starting with 6e77da49e08522f38de4edb2e3b6b126fd4614e853551ad01e64cf5b0a938e65 not found: ID does not exist" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.433103 4799 scope.go:117] "RemoveContainer" containerID="e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55" Sep 30 14:23:36 crc kubenswrapper[4799]: E0930 14:23:36.433361 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55\": container with ID starting with e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55 not found: ID does not exist" containerID="e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.433389 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55"} err="failed to get container status \"e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55\": rpc error: code = NotFound desc = could not find container \"e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55\": container with ID starting with e11928c3be8eb9f2eef0aec5f4e5e0d11dd854342d30f2f6b497732b25063a55 not found: ID does not exist" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.433416 4799 scope.go:117] "RemoveContainer" containerID="4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24" Sep 30 14:23:36 crc kubenswrapper[4799]: E0930 14:23:36.433680 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24\": container with ID starting with 4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24 not found: ID does not exist" containerID="4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.433705 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24"} err="failed to get container status \"4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24\": rpc error: code = NotFound desc = could not find container \"4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24\": container with ID starting with 4f2e207a23387554302915bb6da24654b4debd55a986321d88427f5ac62f9f24 not found: ID does not exist" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.510226 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" path="/var/lib/kubelet/pods/5ef93211-55a5-43d0-9bda-a1e41c44fb35/volumes" Sep 30 14:23:36 crc kubenswrapper[4799]: I0930 14:23:36.510951 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" path="/var/lib/kubelet/pods/b9eab5bb-9ccf-4184-a82e-25b41844ea1b/volumes" Sep 30 14:23:51 crc kubenswrapper[4799]: I0930 14:23:51.916151 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" podUID="c35fadda-30b5-438e-b0a6-945004f36333" containerName="oauth-openshift" containerID="cri-o://80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6" gracePeriod=15 Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.295041 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.326805 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-rs5dr"] Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.326993 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerName="extract-utilities" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327004 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerName="extract-utilities" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327013 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerName="extract-content" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327019 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerName="extract-content" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327025 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327031 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327040 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327046 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327053 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="extract-content" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327059 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="extract-content" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327066 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327072 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327078 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerName="extract-utilities" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327084 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerName="extract-utilities" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327091 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerName="extract-utilities" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327097 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerName="extract-utilities" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327103 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0" containerName="pruner" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327109 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0" containerName="pruner" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327115 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327121 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327132 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec017ab4-5210-40d6-a33a-ca7ace14bafe" containerName="pruner" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327137 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec017ab4-5210-40d6-a33a-ca7ace14bafe" containerName="pruner" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327146 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35fadda-30b5-438e-b0a6-945004f36333" containerName="oauth-openshift" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327152 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35fadda-30b5-438e-b0a6-945004f36333" containerName="oauth-openshift" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327159 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f147ef4-6fba-40b2-bdda-f19618d512eb" containerName="collect-profiles" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327165 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f147ef4-6fba-40b2-bdda-f19618d512eb" containerName="collect-profiles" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327174 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerName="extract-content" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327179 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerName="extract-content" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327189 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="extract-utilities" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327195 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="extract-utilities" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.327202 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerName="extract-content" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327207 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerName="extract-content" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327288 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc5a4dc6-7585-4682-b51b-65b6c5c7b8b0" containerName="pruner" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327306 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f147ef4-6fba-40b2-bdda-f19618d512eb" containerName="collect-profiles" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327313 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35fadda-30b5-438e-b0a6-945004f36333" containerName="oauth-openshift" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327321 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef93211-55a5-43d0-9bda-a1e41c44fb35" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327329 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="70b0c843-7e0f-4cca-8443-9306f0c7ad76" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327337 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9eab5bb-9ccf-4184-a82e-25b41844ea1b" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327345 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec017ab4-5210-40d6-a33a-ca7ace14bafe" containerName="pruner" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327351 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9dfdd6-a206-4a32-8e66-4402828d4d6c" containerName="registry-server" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.327730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.352015 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-rs5dr"] Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.414807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-trusted-ca-bundle\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.415761 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.415974 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c35fadda-30b5-438e-b0a6-945004f36333-audit-dir\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.416070 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c35fadda-30b5-438e-b0a6-945004f36333-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.416224 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-idp-0-file-data\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.416306 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-ocp-branding-template\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.416398 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-serving-cert\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.416478 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-error\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.416561 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-router-certs\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.416662 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-session\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.416748 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc8pm\" (UniqueName: \"kubernetes.io/projected/c35fadda-30b5-438e-b0a6-945004f36333-kube-api-access-jc8pm\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.417492 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-login\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.417582 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-service-ca\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.417674 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-audit-policies\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.417847 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-provider-selection\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.417936 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-cliconfig\") pod \"c35fadda-30b5-438e-b0a6-945004f36333\" (UID: \"c35fadda-30b5-438e-b0a6-945004f36333\") " Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418157 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47193284-80b0-4937-980a-7fa027239237-audit-dir\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418264 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-audit-policies\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418362 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418462 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418544 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418610 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418795 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418888 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.418962 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.419038 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.419126 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtzr2\" (UniqueName: \"kubernetes.io/projected/47193284-80b0-4937-980a-7fa027239237-kube-api-access-gtzr2\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.419212 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.421564 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.421693 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.421780 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.421880 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.421895 4799 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c35fadda-30b5-438e-b0a6-945004f36333-audit-dir\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.421910 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.422288 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.422670 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.422878 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.423191 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.424124 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.424248 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.424455 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.424826 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.425016 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.431048 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.431915 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c35fadda-30b5-438e-b0a6-945004f36333-kube-api-access-jc8pm" (OuterVolumeSpecName: "kube-api-access-jc8pm") pod "c35fadda-30b5-438e-b0a6-945004f36333" (UID: "c35fadda-30b5-438e-b0a6-945004f36333"). InnerVolumeSpecName "kube-api-access-jc8pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.476134 4799 generic.go:334] "Generic (PLEG): container finished" podID="c35fadda-30b5-438e-b0a6-945004f36333" containerID="80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6" exitCode=0 Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.476175 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" event={"ID":"c35fadda-30b5-438e-b0a6-945004f36333","Type":"ContainerDied","Data":"80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6"} Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.476201 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" event={"ID":"c35fadda-30b5-438e-b0a6-945004f36333","Type":"ContainerDied","Data":"14779a4953735ff9ef866ce4d6a55adaead399af35abcfeb91167287895d7c4d"} Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.476219 4799 scope.go:117] "RemoveContainer" containerID="80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.476337 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfxv2" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.495937 4799 scope.go:117] "RemoveContainer" containerID="80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6" Sep 30 14:23:52 crc kubenswrapper[4799]: E0930 14:23:52.496963 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6\": container with ID starting with 80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6 not found: ID does not exist" containerID="80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.497005 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6"} err="failed to get container status \"80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6\": rpc error: code = NotFound desc = could not find container \"80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6\": container with ID starting with 80c502ce38275d25508b85c74d3e4d229d681c6fa952c919716c7f802bae5db6 not found: ID does not exist" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.515755 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfxv2"] Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.515790 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfxv2"] Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522562 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-audit-policies\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522594 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522613 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522636 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522688 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522705 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522725 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522746 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522762 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522783 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtzr2\" (UniqueName: \"kubernetes.io/projected/47193284-80b0-4937-980a-7fa027239237-kube-api-access-gtzr2\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522806 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522825 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522849 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522874 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47193284-80b0-4937-980a-7fa027239237-audit-dir\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.522960 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.525315 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-cliconfig\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.525354 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47193284-80b0-4937-980a-7fa027239237-audit-dir\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.525544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-service-ca\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.525686 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-audit-policies\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.526009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.527394 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-error\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.527444 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.527456 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.527774 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-login\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528355 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528409 4799 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528425 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528440 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528453 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528468 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528480 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528493 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528505 4799 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c35fadda-30b5-438e-b0a6-945004f36333-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.528518 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc8pm\" (UniqueName: \"kubernetes.io/projected/c35fadda-30b5-438e-b0a6-945004f36333-kube-api-access-jc8pm\") on node \"crc\" DevicePath \"\"" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.529101 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-router-certs\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.529265 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-serving-cert\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.529425 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.529835 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/47193284-80b0-4937-980a-7fa027239237-v4-0-config-system-session\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.540034 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtzr2\" (UniqueName: \"kubernetes.io/projected/47193284-80b0-4937-980a-7fa027239237-kube-api-access-gtzr2\") pod \"oauth-openshift-594b4c596d-rs5dr\" (UID: \"47193284-80b0-4937-980a-7fa027239237\") " pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:52 crc kubenswrapper[4799]: I0930 14:23:52.652320 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:53 crc kubenswrapper[4799]: I0930 14:23:53.066463 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-594b4c596d-rs5dr"] Sep 30 14:23:53 crc kubenswrapper[4799]: I0930 14:23:53.485739 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" event={"ID":"47193284-80b0-4937-980a-7fa027239237","Type":"ContainerStarted","Data":"6c16a019337a5e32ce932a34e0af19661a1f53089dfdbe086b4fdf8c29f503b2"} Sep 30 14:23:53 crc kubenswrapper[4799]: I0930 14:23:53.486020 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" event={"ID":"47193284-80b0-4937-980a-7fa027239237","Type":"ContainerStarted","Data":"0ab0f2fd9a5b4e0c0996c1f83dae3f829f29752875cce395762aafb7d33cd775"} Sep 30 14:23:53 crc kubenswrapper[4799]: I0930 14:23:53.486717 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:53 crc kubenswrapper[4799]: I0930 14:23:53.507332 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" podStartSLOduration=27.507309898 podStartE2EDuration="27.507309898s" podCreationTimestamp="2025-09-30 14:23:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:23:53.504454592 +0000 UTC m=+255.588055039" watchObservedRunningTime="2025-09-30 14:23:53.507309898 +0000 UTC m=+255.590910325" Sep 30 14:23:53 crc kubenswrapper[4799]: I0930 14:23:53.675598 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-594b4c596d-rs5dr" Sep 30 14:23:54 crc kubenswrapper[4799]: I0930 14:23:54.514753 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c35fadda-30b5-438e-b0a6-945004f36333" path="/var/lib/kubelet/pods/c35fadda-30b5-438e-b0a6-945004f36333/volumes" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.094484 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nzj2q"] Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.095547 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nzj2q" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerName="registry-server" containerID="cri-o://6ee699983927862c3a98c1a550fc8f69384bd31cfe40f5202e623df00660f923" gracePeriod=30 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.106828 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76bw2"] Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.107800 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-76bw2" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerName="registry-server" containerID="cri-o://1c94223e7ae062e9c81bc1e246eefe005753ef17edc4fde6d1ea9ca233b9f2be" gracePeriod=30 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.122556 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-b7pv9"] Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.122885 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" podUID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" containerName="marketplace-operator" containerID="cri-o://1297a8efb779e2b6a9969b9b65bf94b81fc99acaa86acbcab71316f4bccde0af" gracePeriod=30 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.137631 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pzj2v"] Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.137968 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pzj2v" podUID="fe8c96af-648f-407f-a61b-d09601743327" containerName="registry-server" containerID="cri-o://340fb776cf7a8f5797e0ff4b269d1eafe1708022ea5de25b9b1c85e6f58062a0" gracePeriod=30 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.158959 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z2qlr"] Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.159875 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.191712 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ntf96"] Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.192010 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ntf96" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerName="registry-server" containerID="cri-o://bce9dcbc749673a2934143c43187441ae7361083f1ae3238810d3e41cb56c974" gracePeriod=30 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.208634 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z2qlr"] Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.232208 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7f71a790-5180-4e2c-be45-837154fa1d62-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.232507 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f71a790-5180-4e2c-be45-837154fa1d62-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.232823 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtsfd\" (UniqueName: \"kubernetes.io/projected/7f71a790-5180-4e2c-be45-837154fa1d62-kube-api-access-xtsfd\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.334784 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtsfd\" (UniqueName: \"kubernetes.io/projected/7f71a790-5180-4e2c-be45-837154fa1d62-kube-api-access-xtsfd\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.334921 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7f71a790-5180-4e2c-be45-837154fa1d62-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.334954 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f71a790-5180-4e2c-be45-837154fa1d62-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.339824 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f71a790-5180-4e2c-be45-837154fa1d62-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.345705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7f71a790-5180-4e2c-be45-837154fa1d62-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.371310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtsfd\" (UniqueName: \"kubernetes.io/projected/7f71a790-5180-4e2c-be45-837154fa1d62-kube-api-access-xtsfd\") pod \"marketplace-operator-79b997595-z2qlr\" (UID: \"7f71a790-5180-4e2c-be45-837154fa1d62\") " pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.479877 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.598572 4799 generic.go:334] "Generic (PLEG): container finished" podID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" containerID="1297a8efb779e2b6a9969b9b65bf94b81fc99acaa86acbcab71316f4bccde0af" exitCode=0 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.598619 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" event={"ID":"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd","Type":"ContainerDied","Data":"1297a8efb779e2b6a9969b9b65bf94b81fc99acaa86acbcab71316f4bccde0af"} Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.602725 4799 generic.go:334] "Generic (PLEG): container finished" podID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerID="bce9dcbc749673a2934143c43187441ae7361083f1ae3238810d3e41cb56c974" exitCode=0 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.602790 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntf96" event={"ID":"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f","Type":"ContainerDied","Data":"bce9dcbc749673a2934143c43187441ae7361083f1ae3238810d3e41cb56c974"} Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.606907 4799 generic.go:334] "Generic (PLEG): container finished" podID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerID="6ee699983927862c3a98c1a550fc8f69384bd31cfe40f5202e623df00660f923" exitCode=0 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.606986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2q" event={"ID":"674513e3-aa0c-4814-884c-d0cb1468a3c9","Type":"ContainerDied","Data":"6ee699983927862c3a98c1a550fc8f69384bd31cfe40f5202e623df00660f923"} Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.623117 4799 generic.go:334] "Generic (PLEG): container finished" podID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerID="1c94223e7ae062e9c81bc1e246eefe005753ef17edc4fde6d1ea9ca233b9f2be" exitCode=0 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.623195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76bw2" event={"ID":"4eb004e8-ef9d-404c-9321-2c566b86e393","Type":"ContainerDied","Data":"1c94223e7ae062e9c81bc1e246eefe005753ef17edc4fde6d1ea9ca233b9f2be"} Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.641453 4799 generic.go:334] "Generic (PLEG): container finished" podID="fe8c96af-648f-407f-a61b-d09601743327" containerID="340fb776cf7a8f5797e0ff4b269d1eafe1708022ea5de25b9b1c85e6f58062a0" exitCode=0 Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.641576 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pzj2v" event={"ID":"fe8c96af-648f-407f-a61b-d09601743327","Type":"ContainerDied","Data":"340fb776cf7a8f5797e0ff4b269d1eafe1708022ea5de25b9b1c85e6f58062a0"} Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.689203 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.751449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-trusted-ca\") pod \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.751606 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5rcq\" (UniqueName: \"kubernetes.io/projected/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-kube-api-access-k5rcq\") pod \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.751711 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-operator-metrics\") pod \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\" (UID: \"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd\") " Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.753331 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" (UID: "b3e7cfc7-e3e1-43a3-9880-a628fc0afafd"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.756297 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" (UID: "b3e7cfc7-e3e1-43a3-9880-a628fc0afafd"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.757355 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-kube-api-access-k5rcq" (OuterVolumeSpecName: "kube-api-access-k5rcq") pod "b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" (UID: "b3e7cfc7-e3e1-43a3-9880-a628fc0afafd"). InnerVolumeSpecName "kube-api-access-k5rcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.854345 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.854385 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5rcq\" (UniqueName: \"kubernetes.io/projected/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-kube-api-access-k5rcq\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.854397 4799 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.926939 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:24:06 crc kubenswrapper[4799]: I0930 14:24:06.989088 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.056975 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-utilities\") pod \"fe8c96af-648f-407f-a61b-d09601743327\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.057360 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-catalog-content\") pod \"4eb004e8-ef9d-404c-9321-2c566b86e393\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.057396 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgh5n\" (UniqueName: \"kubernetes.io/projected/fe8c96af-648f-407f-a61b-d09601743327-kube-api-access-lgh5n\") pod \"fe8c96af-648f-407f-a61b-d09601743327\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.057443 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-utilities\") pod \"4eb004e8-ef9d-404c-9321-2c566b86e393\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.057484 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5f8w\" (UniqueName: \"kubernetes.io/projected/4eb004e8-ef9d-404c-9321-2c566b86e393-kube-api-access-z5f8w\") pod \"4eb004e8-ef9d-404c-9321-2c566b86e393\" (UID: \"4eb004e8-ef9d-404c-9321-2c566b86e393\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.057519 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-catalog-content\") pod \"fe8c96af-648f-407f-a61b-d09601743327\" (UID: \"fe8c96af-648f-407f-a61b-d09601743327\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.058387 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-utilities" (OuterVolumeSpecName: "utilities") pod "4eb004e8-ef9d-404c-9321-2c566b86e393" (UID: "4eb004e8-ef9d-404c-9321-2c566b86e393"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.058988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-utilities" (OuterVolumeSpecName: "utilities") pod "fe8c96af-648f-407f-a61b-d09601743327" (UID: "fe8c96af-648f-407f-a61b-d09601743327"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.066892 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.096936 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe8c96af-648f-407f-a61b-d09601743327-kube-api-access-lgh5n" (OuterVolumeSpecName: "kube-api-access-lgh5n") pod "fe8c96af-648f-407f-a61b-d09601743327" (UID: "fe8c96af-648f-407f-a61b-d09601743327"). InnerVolumeSpecName "kube-api-access-lgh5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.101713 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eb004e8-ef9d-404c-9321-2c566b86e393-kube-api-access-z5f8w" (OuterVolumeSpecName: "kube-api-access-z5f8w") pod "4eb004e8-ef9d-404c-9321-2c566b86e393" (UID: "4eb004e8-ef9d-404c-9321-2c566b86e393"). InnerVolumeSpecName "kube-api-access-z5f8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.106643 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe8c96af-648f-407f-a61b-d09601743327" (UID: "fe8c96af-648f-407f-a61b-d09601743327"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.128229 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.158534 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnslj\" (UniqueName: \"kubernetes.io/projected/674513e3-aa0c-4814-884c-d0cb1468a3c9-kube-api-access-jnslj\") pod \"674513e3-aa0c-4814-884c-d0cb1468a3c9\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.158583 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-catalog-content\") pod \"674513e3-aa0c-4814-884c-d0cb1468a3c9\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.158750 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-utilities\") pod \"674513e3-aa0c-4814-884c-d0cb1468a3c9\" (UID: \"674513e3-aa0c-4814-884c-d0cb1468a3c9\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.158978 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5f8w\" (UniqueName: \"kubernetes.io/projected/4eb004e8-ef9d-404c-9321-2c566b86e393-kube-api-access-z5f8w\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.158999 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.159014 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe8c96af-648f-407f-a61b-d09601743327-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.159026 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgh5n\" (UniqueName: \"kubernetes.io/projected/fe8c96af-648f-407f-a61b-d09601743327-kube-api-access-lgh5n\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.159038 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.160476 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-utilities" (OuterVolumeSpecName: "utilities") pod "674513e3-aa0c-4814-884c-d0cb1468a3c9" (UID: "674513e3-aa0c-4814-884c-d0cb1468a3c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.176693 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/674513e3-aa0c-4814-884c-d0cb1468a3c9-kube-api-access-jnslj" (OuterVolumeSpecName: "kube-api-access-jnslj") pod "674513e3-aa0c-4814-884c-d0cb1468a3c9" (UID: "674513e3-aa0c-4814-884c-d0cb1468a3c9"). InnerVolumeSpecName "kube-api-access-jnslj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.178614 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4eb004e8-ef9d-404c-9321-2c566b86e393" (UID: "4eb004e8-ef9d-404c-9321-2c566b86e393"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.212037 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "674513e3-aa0c-4814-884c-d0cb1468a3c9" (UID: "674513e3-aa0c-4814-884c-d0cb1468a3c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.259396 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7vqt\" (UniqueName: \"kubernetes.io/projected/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-kube-api-access-j7vqt\") pod \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.259478 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-catalog-content\") pod \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.259559 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-utilities\") pod \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\" (UID: \"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f\") " Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.259833 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.259851 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb004e8-ef9d-404c-9321-2c566b86e393-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.259862 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnslj\" (UniqueName: \"kubernetes.io/projected/674513e3-aa0c-4814-884c-d0cb1468a3c9-kube-api-access-jnslj\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.259871 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/674513e3-aa0c-4814-884c-d0cb1468a3c9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.260829 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-utilities" (OuterVolumeSpecName: "utilities") pod "d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" (UID: "d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.263262 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-kube-api-access-j7vqt" (OuterVolumeSpecName: "kube-api-access-j7vqt") pod "d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" (UID: "d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f"). InnerVolumeSpecName "kube-api-access-j7vqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.309946 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z2qlr"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.361131 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.361179 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7vqt\" (UniqueName: \"kubernetes.io/projected/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-kube-api-access-j7vqt\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.386625 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" (UID: "d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.462749 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.649624 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pzj2v" event={"ID":"fe8c96af-648f-407f-a61b-d09601743327","Type":"ContainerDied","Data":"f222c0d2e8828998b9250abca65a28f6fa91136ab1db51acd6e284936f504879"} Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.649737 4799 scope.go:117] "RemoveContainer" containerID="340fb776cf7a8f5797e0ff4b269d1eafe1708022ea5de25b9b1c85e6f58062a0" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.650209 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pzj2v" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.651292 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" event={"ID":"7f71a790-5180-4e2c-be45-837154fa1d62","Type":"ContainerStarted","Data":"d8f39e8e2eb86f53bd208af45fedfd638722162fe13d62301d4ad84ba4f036e2"} Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.651326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" event={"ID":"7f71a790-5180-4e2c-be45-837154fa1d62","Type":"ContainerStarted","Data":"9fb835bd9cb87eecc19efcc79101f7f0759881e528c3f632b3f795008ae3e402"} Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.651537 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.653044 4799 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z2qlr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.653090 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" podUID="7f71a790-5180-4e2c-be45-837154fa1d62" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.653140 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" event={"ID":"b3e7cfc7-e3e1-43a3-9880-a628fc0afafd","Type":"ContainerDied","Data":"8825ef73d3b0f29c2278b96c6dc0bed273a89c90152cff23c5ebb64222ff7f29"} Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.653167 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-b7pv9" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.655887 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntf96" event={"ID":"d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f","Type":"ContainerDied","Data":"3656e658885737a4776fc63482390307c4b6c0b8197c41ea78faa459ab982726"} Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.655982 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntf96" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.662751 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nzj2q" event={"ID":"674513e3-aa0c-4814-884c-d0cb1468a3c9","Type":"ContainerDied","Data":"af4f891e954a5b8041b1baec0b3b8d70853600f1b2733e851e39241c23b1bb4b"} Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.662928 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nzj2q" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.665760 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76bw2" event={"ID":"4eb004e8-ef9d-404c-9321-2c566b86e393","Type":"ContainerDied","Data":"742e75b947f7e922960ba5edcbbdf5deb50ec555354d4f52fe1ae4aef0041101"} Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.665879 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76bw2" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.676996 4799 scope.go:117] "RemoveContainer" containerID="5e5d59325a73fdfa68a2e35c7f23840faf3c6611e663d3124b07e5cc491b99b6" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.684913 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" podStartSLOduration=1.684895081 podStartE2EDuration="1.684895081s" podCreationTimestamp="2025-09-30 14:24:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:24:07.682922461 +0000 UTC m=+269.766522908" watchObservedRunningTime="2025-09-30 14:24:07.684895081 +0000 UTC m=+269.768495508" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.711797 4799 scope.go:117] "RemoveContainer" containerID="373648fdc330acb0b2c9c2831cf7f0335c99a6d524a04d87be02842974830083" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.736626 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-b7pv9"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.739976 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-b7pv9"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.752633 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ntf96"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.755838 4799 scope.go:117] "RemoveContainer" containerID="1297a8efb779e2b6a9969b9b65bf94b81fc99acaa86acbcab71316f4bccde0af" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.756463 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ntf96"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.767539 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76bw2"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.770470 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-76bw2"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.776733 4799 scope.go:117] "RemoveContainer" containerID="bce9dcbc749673a2934143c43187441ae7361083f1ae3238810d3e41cb56c974" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.779389 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nzj2q"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.785017 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nzj2q"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.799067 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pzj2v"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.802108 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pzj2v"] Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.802988 4799 scope.go:117] "RemoveContainer" containerID="6e49d00cb0f35ee40f271b85e9d53c4ab7309818d02d7f31ab65426006d6b626" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.818275 4799 scope.go:117] "RemoveContainer" containerID="b66989d24033687a618f63d84920f67f78b92e148b29eb50e127aac162ccb313" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.836404 4799 scope.go:117] "RemoveContainer" containerID="6ee699983927862c3a98c1a550fc8f69384bd31cfe40f5202e623df00660f923" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.854411 4799 scope.go:117] "RemoveContainer" containerID="765af5ef5a37b37e375b12aaa903bb8e88a708bc80e9d704d00839c35c9de38f" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.868881 4799 scope.go:117] "RemoveContainer" containerID="6880bf8456dfd4e03ded124a780e099294a4a83057fd6e7c678eba615f5fc265" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.891261 4799 scope.go:117] "RemoveContainer" containerID="1c94223e7ae062e9c81bc1e246eefe005753ef17edc4fde6d1ea9ca233b9f2be" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.907944 4799 scope.go:117] "RemoveContainer" containerID="c7e5ac575f406dbc52a27b109749f519eb14c02a6f77c06db747b0ff3c8024aa" Sep 30 14:24:07 crc kubenswrapper[4799]: I0930 14:24:07.932159 4799 scope.go:117] "RemoveContainer" containerID="b29181a6b5ca736b39b9963fe174acf631f3cac125731f5e50548edb69f593bc" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315415 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5d6jh"] Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315665 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerName="extract-utilities" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315680 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerName="extract-utilities" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315695 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315703 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315716 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315723 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315735 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe8c96af-648f-407f-a61b-d09601743327" containerName="extract-content" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315744 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe8c96af-648f-407f-a61b-d09601743327" containerName="extract-content" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315752 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerName="extract-content" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315758 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerName="extract-content" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315768 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerName="extract-content" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315775 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerName="extract-content" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315782 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe8c96af-648f-407f-a61b-d09601743327" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315789 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe8c96af-648f-407f-a61b-d09601743327" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315801 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe8c96af-648f-407f-a61b-d09601743327" containerName="extract-utilities" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315807 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe8c96af-648f-407f-a61b-d09601743327" containerName="extract-utilities" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315816 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" containerName="marketplace-operator" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315823 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" containerName="marketplace-operator" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315831 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerName="extract-content" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315837 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerName="extract-content" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315845 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerName="extract-utilities" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315853 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerName="extract-utilities" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315863 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315870 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: E0930 14:24:08.315885 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerName="extract-utilities" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315891 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerName="extract-utilities" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315988 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.315998 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.316008 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe8c96af-648f-407f-a61b-d09601743327" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.316018 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" containerName="registry-server" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.316026 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" containerName="marketplace-operator" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.316871 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.319171 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.336615 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5d6jh"] Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.372916 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171994f2-bcb6-4253-a9c1-bfac1e7b1332-catalog-content\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.373028 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szltt\" (UniqueName: \"kubernetes.io/projected/171994f2-bcb6-4253-a9c1-bfac1e7b1332-kube-api-access-szltt\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.373057 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171994f2-bcb6-4253-a9c1-bfac1e7b1332-utilities\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.474086 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171994f2-bcb6-4253-a9c1-bfac1e7b1332-catalog-content\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.474163 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szltt\" (UniqueName: \"kubernetes.io/projected/171994f2-bcb6-4253-a9c1-bfac1e7b1332-kube-api-access-szltt\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.474293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171994f2-bcb6-4253-a9c1-bfac1e7b1332-utilities\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.474818 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171994f2-bcb6-4253-a9c1-bfac1e7b1332-catalog-content\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.475042 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171994f2-bcb6-4253-a9c1-bfac1e7b1332-utilities\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.492978 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szltt\" (UniqueName: \"kubernetes.io/projected/171994f2-bcb6-4253-a9c1-bfac1e7b1332-kube-api-access-szltt\") pod \"certified-operators-5d6jh\" (UID: \"171994f2-bcb6-4253-a9c1-bfac1e7b1332\") " pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.512798 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eb004e8-ef9d-404c-9321-2c566b86e393" path="/var/lib/kubelet/pods/4eb004e8-ef9d-404c-9321-2c566b86e393/volumes" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.513541 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="674513e3-aa0c-4814-884c-d0cb1468a3c9" path="/var/lib/kubelet/pods/674513e3-aa0c-4814-884c-d0cb1468a3c9/volumes" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.514315 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e7cfc7-e3e1-43a3-9880-a628fc0afafd" path="/var/lib/kubelet/pods/b3e7cfc7-e3e1-43a3-9880-a628fc0afafd/volumes" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.515625 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f" path="/var/lib/kubelet/pods/d40d4b92-2ec5-41b5-bb7d-0c49caf82e6f/volumes" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.516512 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe8c96af-648f-407f-a61b-d09601743327" path="/var/lib/kubelet/pods/fe8c96af-648f-407f-a61b-d09601743327/volumes" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.517151 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rmb2k"] Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.521908 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.529397 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.532235 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmb2k"] Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.575313 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fclz4\" (UniqueName: \"kubernetes.io/projected/edfa00ec-c706-4129-b0d2-baf76f31b44c-kube-api-access-fclz4\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.575706 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edfa00ec-c706-4129-b0d2-baf76f31b44c-catalog-content\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.575846 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edfa00ec-c706-4129-b0d2-baf76f31b44c-utilities\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.644634 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.677184 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fclz4\" (UniqueName: \"kubernetes.io/projected/edfa00ec-c706-4129-b0d2-baf76f31b44c-kube-api-access-fclz4\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.677499 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edfa00ec-c706-4129-b0d2-baf76f31b44c-catalog-content\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.677663 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edfa00ec-c706-4129-b0d2-baf76f31b44c-utilities\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.678354 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edfa00ec-c706-4129-b0d2-baf76f31b44c-catalog-content\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.678439 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edfa00ec-c706-4129-b0d2-baf76f31b44c-utilities\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.701147 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fclz4\" (UniqueName: \"kubernetes.io/projected/edfa00ec-c706-4129-b0d2-baf76f31b44c-kube-api-access-fclz4\") pod \"redhat-marketplace-rmb2k\" (UID: \"edfa00ec-c706-4129-b0d2-baf76f31b44c\") " pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.704813 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z2qlr" Sep 30 14:24:08 crc kubenswrapper[4799]: I0930 14:24:08.850096 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:09 crc kubenswrapper[4799]: I0930 14:24:09.117085 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5d6jh"] Sep 30 14:24:09 crc kubenswrapper[4799]: W0930 14:24:09.125917 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod171994f2_bcb6_4253_a9c1_bfac1e7b1332.slice/crio-2ac7d7b33c6b9e355a67e4dee2e9ba040d69b557c31dc9992782c9d8e6dcb34f WatchSource:0}: Error finding container 2ac7d7b33c6b9e355a67e4dee2e9ba040d69b557c31dc9992782c9d8e6dcb34f: Status 404 returned error can't find the container with id 2ac7d7b33c6b9e355a67e4dee2e9ba040d69b557c31dc9992782c9d8e6dcb34f Sep 30 14:24:09 crc kubenswrapper[4799]: I0930 14:24:09.285265 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmb2k"] Sep 30 14:24:09 crc kubenswrapper[4799]: W0930 14:24:09.295259 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedfa00ec_c706_4129_b0d2_baf76f31b44c.slice/crio-06c8a814498d6fbec19d2379fbdfe8c00030f2e00fda4d7f57da2a08144b9b80 WatchSource:0}: Error finding container 06c8a814498d6fbec19d2379fbdfe8c00030f2e00fda4d7f57da2a08144b9b80: Status 404 returned error can't find the container with id 06c8a814498d6fbec19d2379fbdfe8c00030f2e00fda4d7f57da2a08144b9b80 Sep 30 14:24:09 crc kubenswrapper[4799]: I0930 14:24:09.709361 4799 generic.go:334] "Generic (PLEG): container finished" podID="171994f2-bcb6-4253-a9c1-bfac1e7b1332" containerID="d21c10e90b711deeb723a97c82d08ada2e49514598b12e2f64fe42fcde9e3fee" exitCode=0 Sep 30 14:24:09 crc kubenswrapper[4799]: I0930 14:24:09.709433 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d6jh" event={"ID":"171994f2-bcb6-4253-a9c1-bfac1e7b1332","Type":"ContainerDied","Data":"d21c10e90b711deeb723a97c82d08ada2e49514598b12e2f64fe42fcde9e3fee"} Sep 30 14:24:09 crc kubenswrapper[4799]: I0930 14:24:09.709465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d6jh" event={"ID":"171994f2-bcb6-4253-a9c1-bfac1e7b1332","Type":"ContainerStarted","Data":"2ac7d7b33c6b9e355a67e4dee2e9ba040d69b557c31dc9992782c9d8e6dcb34f"} Sep 30 14:24:09 crc kubenswrapper[4799]: I0930 14:24:09.712501 4799 generic.go:334] "Generic (PLEG): container finished" podID="edfa00ec-c706-4129-b0d2-baf76f31b44c" containerID="2cfed9876bea4c6cce5836d7d29d1e4c3f0c7f7c9a93c70db8ad8a0e2fb395ea" exitCode=0 Sep 30 14:24:09 crc kubenswrapper[4799]: I0930 14:24:09.712791 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmb2k" event={"ID":"edfa00ec-c706-4129-b0d2-baf76f31b44c","Type":"ContainerDied","Data":"2cfed9876bea4c6cce5836d7d29d1e4c3f0c7f7c9a93c70db8ad8a0e2fb395ea"} Sep 30 14:24:09 crc kubenswrapper[4799]: I0930 14:24:09.712841 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmb2k" event={"ID":"edfa00ec-c706-4129-b0d2-baf76f31b44c","Type":"ContainerStarted","Data":"06c8a814498d6fbec19d2379fbdfe8c00030f2e00fda4d7f57da2a08144b9b80"} Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.727049 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6nbzj"] Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.728070 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.734719 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.737137 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmb2k" event={"ID":"edfa00ec-c706-4129-b0d2-baf76f31b44c","Type":"ContainerStarted","Data":"1d45b8fc9b7c6c7bf177f17402caaf59e24200be5059728a423607f6ce8556b5"} Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.760914 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nbzj"] Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.809998 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-catalog-content\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.810045 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pghhq\" (UniqueName: \"kubernetes.io/projected/7635e2e9-575c-43ff-894f-e8bd89dacb28-kube-api-access-pghhq\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.810081 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-utilities\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.911717 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-catalog-content\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.912284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pghhq\" (UniqueName: \"kubernetes.io/projected/7635e2e9-575c-43ff-894f-e8bd89dacb28-kube-api-access-pghhq\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.912321 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-utilities\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.912522 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-catalog-content\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.912712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-utilities\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.916141 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tplx2"] Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.919122 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.923336 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.932629 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tplx2"] Sep 30 14:24:10 crc kubenswrapper[4799]: I0930 14:24:10.957871 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pghhq\" (UniqueName: \"kubernetes.io/projected/7635e2e9-575c-43ff-894f-e8bd89dacb28-kube-api-access-pghhq\") pod \"community-operators-6nbzj\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.013805 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8gm4\" (UniqueName: \"kubernetes.io/projected/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-kube-api-access-r8gm4\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.013989 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-utilities\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.014025 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-catalog-content\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.080821 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.115552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-utilities\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.115614 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-catalog-content\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.115643 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8gm4\" (UniqueName: \"kubernetes.io/projected/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-kube-api-access-r8gm4\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.116233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-utilities\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.116820 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-catalog-content\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.134488 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8gm4\" (UniqueName: \"kubernetes.io/projected/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-kube-api-access-r8gm4\") pod \"redhat-operators-tplx2\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.240501 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.515714 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nbzj"] Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.667093 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tplx2"] Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.745578 4799 generic.go:334] "Generic (PLEG): container finished" podID="edfa00ec-c706-4129-b0d2-baf76f31b44c" containerID="1d45b8fc9b7c6c7bf177f17402caaf59e24200be5059728a423607f6ce8556b5" exitCode=0 Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.745617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmb2k" event={"ID":"edfa00ec-c706-4129-b0d2-baf76f31b44c","Type":"ContainerDied","Data":"1d45b8fc9b7c6c7bf177f17402caaf59e24200be5059728a423607f6ce8556b5"} Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.759549 4799 generic.go:334] "Generic (PLEG): container finished" podID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerID="73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64" exitCode=0 Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.759642 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nbzj" event={"ID":"7635e2e9-575c-43ff-894f-e8bd89dacb28","Type":"ContainerDied","Data":"73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64"} Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.759686 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nbzj" event={"ID":"7635e2e9-575c-43ff-894f-e8bd89dacb28","Type":"ContainerStarted","Data":"7ec2838cb8173e1c7b4038a2d8fbdd2234d70e55d6c268fd5ef476b0b55b068a"} Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.770168 4799 generic.go:334] "Generic (PLEG): container finished" podID="171994f2-bcb6-4253-a9c1-bfac1e7b1332" containerID="3651ea3ecb5e91a381d75f41025eda2e3d558a087e370c3ecdc5fce1dd63052a" exitCode=0 Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.770824 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d6jh" event={"ID":"171994f2-bcb6-4253-a9c1-bfac1e7b1332","Type":"ContainerDied","Data":"3651ea3ecb5e91a381d75f41025eda2e3d558a087e370c3ecdc5fce1dd63052a"} Sep 30 14:24:11 crc kubenswrapper[4799]: I0930 14:24:11.786445 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tplx2" event={"ID":"eb9f79d7-c4db-4f1f-9b99-6c577dca041d","Type":"ContainerStarted","Data":"bf8bfcd40b9a7fd5b501abe6baf79fe640957e9e27ea58b410cd4f9b8be4627b"} Sep 30 14:24:12 crc kubenswrapper[4799]: I0930 14:24:12.795207 4799 generic.go:334] "Generic (PLEG): container finished" podID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerID="2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f" exitCode=0 Sep 30 14:24:12 crc kubenswrapper[4799]: I0930 14:24:12.795439 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tplx2" event={"ID":"eb9f79d7-c4db-4f1f-9b99-6c577dca041d","Type":"ContainerDied","Data":"2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f"} Sep 30 14:24:12 crc kubenswrapper[4799]: I0930 14:24:12.801314 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmb2k" event={"ID":"edfa00ec-c706-4129-b0d2-baf76f31b44c","Type":"ContainerStarted","Data":"8af62aa65c682a4afeda047461392d9fb19261ee481b33128f269888b1fcb142"} Sep 30 14:24:12 crc kubenswrapper[4799]: I0930 14:24:12.841027 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rmb2k" podStartSLOduration=2.431693732 podStartE2EDuration="4.841010704s" podCreationTimestamp="2025-09-30 14:24:08 +0000 UTC" firstStartedPulling="2025-09-30 14:24:09.714318608 +0000 UTC m=+271.797919035" lastFinishedPulling="2025-09-30 14:24:12.12363558 +0000 UTC m=+274.207236007" observedRunningTime="2025-09-30 14:24:12.83594918 +0000 UTC m=+274.919549627" watchObservedRunningTime="2025-09-30 14:24:12.841010704 +0000 UTC m=+274.924611131" Sep 30 14:24:13 crc kubenswrapper[4799]: I0930 14:24:13.819570 4799 generic.go:334] "Generic (PLEG): container finished" podID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerID="dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801" exitCode=0 Sep 30 14:24:13 crc kubenswrapper[4799]: I0930 14:24:13.819667 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nbzj" event={"ID":"7635e2e9-575c-43ff-894f-e8bd89dacb28","Type":"ContainerDied","Data":"dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801"} Sep 30 14:24:13 crc kubenswrapper[4799]: I0930 14:24:13.826877 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d6jh" event={"ID":"171994f2-bcb6-4253-a9c1-bfac1e7b1332","Type":"ContainerStarted","Data":"be9151486bd43157d27913af8c63b94a2cb66236b1e0029a67516be73dded507"} Sep 30 14:24:13 crc kubenswrapper[4799]: I0930 14:24:13.873387 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5d6jh" podStartSLOduration=2.568774015 podStartE2EDuration="5.873362818s" podCreationTimestamp="2025-09-30 14:24:08 +0000 UTC" firstStartedPulling="2025-09-30 14:24:09.71274441 +0000 UTC m=+271.796344837" lastFinishedPulling="2025-09-30 14:24:13.017333213 +0000 UTC m=+275.100933640" observedRunningTime="2025-09-30 14:24:13.86423136 +0000 UTC m=+275.947831787" watchObservedRunningTime="2025-09-30 14:24:13.873362818 +0000 UTC m=+275.956963265" Sep 30 14:24:14 crc kubenswrapper[4799]: I0930 14:24:14.833267 4799 generic.go:334] "Generic (PLEG): container finished" podID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerID="14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb" exitCode=0 Sep 30 14:24:14 crc kubenswrapper[4799]: I0930 14:24:14.833462 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tplx2" event={"ID":"eb9f79d7-c4db-4f1f-9b99-6c577dca041d","Type":"ContainerDied","Data":"14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb"} Sep 30 14:24:16 crc kubenswrapper[4799]: I0930 14:24:16.846735 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nbzj" event={"ID":"7635e2e9-575c-43ff-894f-e8bd89dacb28","Type":"ContainerStarted","Data":"98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f"} Sep 30 14:24:16 crc kubenswrapper[4799]: I0930 14:24:16.849361 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tplx2" event={"ID":"eb9f79d7-c4db-4f1f-9b99-6c577dca041d","Type":"ContainerStarted","Data":"90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d"} Sep 30 14:24:16 crc kubenswrapper[4799]: I0930 14:24:16.869769 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6nbzj" podStartSLOduration=4.29869114 podStartE2EDuration="6.869752187s" podCreationTimestamp="2025-09-30 14:24:10 +0000 UTC" firstStartedPulling="2025-09-30 14:24:11.765528386 +0000 UTC m=+273.849128813" lastFinishedPulling="2025-09-30 14:24:14.336589433 +0000 UTC m=+276.420189860" observedRunningTime="2025-09-30 14:24:16.868460438 +0000 UTC m=+278.952060885" watchObservedRunningTime="2025-09-30 14:24:16.869752187 +0000 UTC m=+278.953352614" Sep 30 14:24:16 crc kubenswrapper[4799]: I0930 14:24:16.888928 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tplx2" podStartSLOduration=4.406515673 podStartE2EDuration="6.88891076s" podCreationTimestamp="2025-09-30 14:24:10 +0000 UTC" firstStartedPulling="2025-09-30 14:24:12.796921282 +0000 UTC m=+274.880521709" lastFinishedPulling="2025-09-30 14:24:15.279316369 +0000 UTC m=+277.362916796" observedRunningTime="2025-09-30 14:24:16.886550759 +0000 UTC m=+278.970151206" watchObservedRunningTime="2025-09-30 14:24:16.88891076 +0000 UTC m=+278.972511187" Sep 30 14:24:18 crc kubenswrapper[4799]: I0930 14:24:18.646003 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:18 crc kubenswrapper[4799]: I0930 14:24:18.646485 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:18 crc kubenswrapper[4799]: I0930 14:24:18.696612 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:18 crc kubenswrapper[4799]: I0930 14:24:18.850548 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:18 crc kubenswrapper[4799]: I0930 14:24:18.850618 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:18 crc kubenswrapper[4799]: I0930 14:24:18.899951 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:18 crc kubenswrapper[4799]: I0930 14:24:18.915866 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5d6jh" Sep 30 14:24:18 crc kubenswrapper[4799]: I0930 14:24:18.951218 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rmb2k" Sep 30 14:24:21 crc kubenswrapper[4799]: I0930 14:24:21.081217 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:21 crc kubenswrapper[4799]: I0930 14:24:21.081789 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:21 crc kubenswrapper[4799]: I0930 14:24:21.131017 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:24:21 crc kubenswrapper[4799]: I0930 14:24:21.241168 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:21 crc kubenswrapper[4799]: I0930 14:24:21.241444 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:21 crc kubenswrapper[4799]: I0930 14:24:21.278534 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:21 crc kubenswrapper[4799]: I0930 14:24:21.912180 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 14:24:21 crc kubenswrapper[4799]: I0930 14:24:21.914218 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6nbzj" Sep 30 14:25:29 crc kubenswrapper[4799]: I0930 14:25:29.648913 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:25:29 crc kubenswrapper[4799]: I0930 14:25:29.650814 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:25:59 crc kubenswrapper[4799]: I0930 14:25:59.648530 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:25:59 crc kubenswrapper[4799]: I0930 14:25:59.649155 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.504511 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-khtt6"] Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.506752 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.533204 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-khtt6"] Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.600073 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d39f74c3-65fb-437d-8f60-d39223711116-ca-trust-extracted\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.600142 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d39f74c3-65fb-437d-8f60-d39223711116-trusted-ca\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.600174 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.600217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d39f74c3-65fb-437d-8f60-d39223711116-installation-pull-secrets\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.600250 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d39f74c3-65fb-437d-8f60-d39223711116-registry-certificates\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.600303 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cmdp\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-kube-api-access-9cmdp\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.600336 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-registry-tls\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.600385 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-bound-sa-token\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.701855 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-bound-sa-token\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.702201 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d39f74c3-65fb-437d-8f60-d39223711116-ca-trust-extracted\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.702292 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d39f74c3-65fb-437d-8f60-d39223711116-trusted-ca\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.702399 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d39f74c3-65fb-437d-8f60-d39223711116-installation-pull-secrets\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.702509 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d39f74c3-65fb-437d-8f60-d39223711116-registry-certificates\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.702627 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cmdp\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-kube-api-access-9cmdp\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.702792 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-registry-tls\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.702858 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d39f74c3-65fb-437d-8f60-d39223711116-ca-trust-extracted\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.703807 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d39f74c3-65fb-437d-8f60-d39223711116-trusted-ca\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.703918 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d39f74c3-65fb-437d-8f60-d39223711116-registry-certificates\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.712438 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-registry-tls\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.716288 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.717829 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d39f74c3-65fb-437d-8f60-d39223711116-installation-pull-secrets\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.721196 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cmdp\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-kube-api-access-9cmdp\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.723054 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d39f74c3-65fb-437d-8f60-d39223711116-bound-sa-token\") pod \"image-registry-66df7c8f76-khtt6\" (UID: \"d39f74c3-65fb-437d-8f60-d39223711116\") " pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:17 crc kubenswrapper[4799]: I0930 14:26:17.824095 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:18 crc kubenswrapper[4799]: I0930 14:26:18.019028 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-khtt6"] Sep 30 14:26:18 crc kubenswrapper[4799]: I0930 14:26:18.496370 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" event={"ID":"d39f74c3-65fb-437d-8f60-d39223711116","Type":"ContainerStarted","Data":"e2edf2f7a83c780d9bab49554c5d53116c5926a0be9024ad9a4d1fbb3b9f1955"} Sep 30 14:26:18 crc kubenswrapper[4799]: I0930 14:26:18.496422 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" event={"ID":"d39f74c3-65fb-437d-8f60-d39223711116","Type":"ContainerStarted","Data":"46e1863ab78880f2cdb3358e5f95c7832ed3daaedd6b3c321db951c7d26dbaa9"} Sep 30 14:26:18 crc kubenswrapper[4799]: I0930 14:26:18.496561 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:18 crc kubenswrapper[4799]: I0930 14:26:18.513901 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" podStartSLOduration=1.5138804430000001 podStartE2EDuration="1.513880443s" podCreationTimestamp="2025-09-30 14:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:26:18.51130242 +0000 UTC m=+400.594902857" watchObservedRunningTime="2025-09-30 14:26:18.513880443 +0000 UTC m=+400.597480870" Sep 30 14:26:29 crc kubenswrapper[4799]: I0930 14:26:29.649173 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:26:29 crc kubenswrapper[4799]: I0930 14:26:29.649929 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:26:29 crc kubenswrapper[4799]: I0930 14:26:29.649994 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:26:29 crc kubenswrapper[4799]: I0930 14:26:29.650869 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"104fd6ad26a2612596b50bf918e18ccbe2581dacca6d978ddfbd4cf3ad4e5322"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:26:29 crc kubenswrapper[4799]: I0930 14:26:29.650940 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://104fd6ad26a2612596b50bf918e18ccbe2581dacca6d978ddfbd4cf3ad4e5322" gracePeriod=600 Sep 30 14:26:30 crc kubenswrapper[4799]: I0930 14:26:30.567246 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="104fd6ad26a2612596b50bf918e18ccbe2581dacca6d978ddfbd4cf3ad4e5322" exitCode=0 Sep 30 14:26:30 crc kubenswrapper[4799]: I0930 14:26:30.567374 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"104fd6ad26a2612596b50bf918e18ccbe2581dacca6d978ddfbd4cf3ad4e5322"} Sep 30 14:26:30 crc kubenswrapper[4799]: I0930 14:26:30.568276 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"d7967cc87bf65a71f9ed110a689be544a6e7e0ea0717c62f31a311bad53a8a0f"} Sep 30 14:26:30 crc kubenswrapper[4799]: I0930 14:26:30.568323 4799 scope.go:117] "RemoveContainer" containerID="70254e2390bde7b4fd22b3a88b31a46a9a5b2c34d90d6ec666d8f6653bd0cbfa" Sep 30 14:26:37 crc kubenswrapper[4799]: I0930 14:26:37.829176 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-khtt6" Sep 30 14:26:37 crc kubenswrapper[4799]: I0930 14:26:37.885174 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pnkgs"] Sep 30 14:27:02 crc kubenswrapper[4799]: I0930 14:27:02.991948 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" podUID="7675ad08-1655-418f-b0c1-baac6a2d35b6" containerName="registry" containerID="cri-o://cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28" gracePeriod=30 Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.313382 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.364258 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-trusted-ca\") pod \"7675ad08-1655-418f-b0c1-baac6a2d35b6\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.364311 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-certificates\") pod \"7675ad08-1655-418f-b0c1-baac6a2d35b6\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.364390 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k779\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-kube-api-access-9k779\") pod \"7675ad08-1655-418f-b0c1-baac6a2d35b6\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.364427 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-bound-sa-token\") pod \"7675ad08-1655-418f-b0c1-baac6a2d35b6\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.364482 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7675ad08-1655-418f-b0c1-baac6a2d35b6-ca-trust-extracted\") pod \"7675ad08-1655-418f-b0c1-baac6a2d35b6\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.364522 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7675ad08-1655-418f-b0c1-baac6a2d35b6-installation-pull-secrets\") pod \"7675ad08-1655-418f-b0c1-baac6a2d35b6\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.364866 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"7675ad08-1655-418f-b0c1-baac6a2d35b6\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.364928 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-tls\") pod \"7675ad08-1655-418f-b0c1-baac6a2d35b6\" (UID: \"7675ad08-1655-418f-b0c1-baac6a2d35b6\") " Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.365447 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "7675ad08-1655-418f-b0c1-baac6a2d35b6" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.367039 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "7675ad08-1655-418f-b0c1-baac6a2d35b6" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.372565 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-kube-api-access-9k779" (OuterVolumeSpecName: "kube-api-access-9k779") pod "7675ad08-1655-418f-b0c1-baac6a2d35b6" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6"). InnerVolumeSpecName "kube-api-access-9k779". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.375368 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7675ad08-1655-418f-b0c1-baac6a2d35b6-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "7675ad08-1655-418f-b0c1-baac6a2d35b6" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.376488 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "7675ad08-1655-418f-b0c1-baac6a2d35b6" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.379694 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "7675ad08-1655-418f-b0c1-baac6a2d35b6" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.381910 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "7675ad08-1655-418f-b0c1-baac6a2d35b6" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.384545 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7675ad08-1655-418f-b0c1-baac6a2d35b6-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "7675ad08-1655-418f-b0c1-baac6a2d35b6" (UID: "7675ad08-1655-418f-b0c1-baac6a2d35b6"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.466572 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k779\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-kube-api-access-9k779\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.466686 4799 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.466717 4799 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7675ad08-1655-418f-b0c1-baac6a2d35b6-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.466738 4799 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7675ad08-1655-418f-b0c1-baac6a2d35b6-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.466758 4799 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.466774 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.466790 4799 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7675ad08-1655-418f-b0c1-baac6a2d35b6-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.743992 4799 generic.go:334] "Generic (PLEG): container finished" podID="7675ad08-1655-418f-b0c1-baac6a2d35b6" containerID="cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28" exitCode=0 Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.744071 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.744093 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" event={"ID":"7675ad08-1655-418f-b0c1-baac6a2d35b6","Type":"ContainerDied","Data":"cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28"} Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.744471 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pnkgs" event={"ID":"7675ad08-1655-418f-b0c1-baac6a2d35b6","Type":"ContainerDied","Data":"9adcc946fb6d41e0e23b0b9ce071785c0ebe473eeef4817dd03f457ceea0c246"} Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.744494 4799 scope.go:117] "RemoveContainer" containerID="cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.782354 4799 scope.go:117] "RemoveContainer" containerID="cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.783400 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pnkgs"] Sep 30 14:27:03 crc kubenswrapper[4799]: E0930 14:27:03.783761 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28\": container with ID starting with cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28 not found: ID does not exist" containerID="cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.783893 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28"} err="failed to get container status \"cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28\": rpc error: code = NotFound desc = could not find container \"cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28\": container with ID starting with cf03b944b9d76a3a04047c20b54d86f42e6ebfb30cdc03c7d8df7a953e88ab28 not found: ID does not exist" Sep 30 14:27:03 crc kubenswrapper[4799]: I0930 14:27:03.788248 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pnkgs"] Sep 30 14:27:04 crc kubenswrapper[4799]: I0930 14:27:04.509601 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7675ad08-1655-418f-b0c1-baac6a2d35b6" path="/var/lib/kubelet/pods/7675ad08-1655-418f-b0c1-baac6a2d35b6/volumes" Sep 30 14:28:29 crc kubenswrapper[4799]: I0930 14:28:29.649371 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:28:29 crc kubenswrapper[4799]: I0930 14:28:29.650822 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:28:38 crc kubenswrapper[4799]: I0930 14:28:38.651017 4799 scope.go:117] "RemoveContainer" containerID="65a963cbbaf84278fe18927cdadee74e0f5be417433733910c961070697ef4eb" Sep 30 14:28:59 crc kubenswrapper[4799]: I0930 14:28:59.649096 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:28:59 crc kubenswrapper[4799]: I0930 14:28:59.649856 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:29:29 crc kubenswrapper[4799]: I0930 14:29:29.649494 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:29:29 crc kubenswrapper[4799]: I0930 14:29:29.650923 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:29:29 crc kubenswrapper[4799]: I0930 14:29:29.650983 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:29:29 crc kubenswrapper[4799]: I0930 14:29:29.651609 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d7967cc87bf65a71f9ed110a689be544a6e7e0ea0717c62f31a311bad53a8a0f"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:29:29 crc kubenswrapper[4799]: I0930 14:29:29.651698 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://d7967cc87bf65a71f9ed110a689be544a6e7e0ea0717c62f31a311bad53a8a0f" gracePeriod=600 Sep 30 14:29:30 crc kubenswrapper[4799]: I0930 14:29:30.519364 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="d7967cc87bf65a71f9ed110a689be544a6e7e0ea0717c62f31a311bad53a8a0f" exitCode=0 Sep 30 14:29:30 crc kubenswrapper[4799]: I0930 14:29:30.519436 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"d7967cc87bf65a71f9ed110a689be544a6e7e0ea0717c62f31a311bad53a8a0f"} Sep 30 14:29:30 crc kubenswrapper[4799]: I0930 14:29:30.519765 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"1e25c11d4641ce442c772191497cb466d8f447d176e97bcb309720e08dd67a66"} Sep 30 14:29:30 crc kubenswrapper[4799]: I0930 14:29:30.519792 4799 scope.go:117] "RemoveContainer" containerID="104fd6ad26a2612596b50bf918e18ccbe2581dacca6d978ddfbd4cf3ad4e5322" Sep 30 14:29:38 crc kubenswrapper[4799]: I0930 14:29:38.685786 4799 scope.go:117] "RemoveContainer" containerID="7edd06435a6f3b27882013d5b3fe1b4c859055bc0fd0ea83d39901ac235b3a04" Sep 30 14:29:38 crc kubenswrapper[4799]: I0930 14:29:38.708637 4799 scope.go:117] "RemoveContainer" containerID="72bc9602034e089cc8b852a788f2a499f2f8b29b1ff6ddecf66cce168dbaef6a" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.140765 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd"] Sep 30 14:30:00 crc kubenswrapper[4799]: E0930 14:30:00.141314 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7675ad08-1655-418f-b0c1-baac6a2d35b6" containerName="registry" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.141333 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7675ad08-1655-418f-b0c1-baac6a2d35b6" containerName="registry" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.141457 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7675ad08-1655-418f-b0c1-baac6a2d35b6" containerName="registry" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.141938 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.143691 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.143998 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.195189 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd"] Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.208597 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5508bbb-b87a-4e2e-bc55-940edda975cc-secret-volume\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.208698 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54klh\" (UniqueName: \"kubernetes.io/projected/a5508bbb-b87a-4e2e-bc55-940edda975cc-kube-api-access-54klh\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.208727 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5508bbb-b87a-4e2e-bc55-940edda975cc-config-volume\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.309840 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5508bbb-b87a-4e2e-bc55-940edda975cc-secret-volume\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.309919 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54klh\" (UniqueName: \"kubernetes.io/projected/a5508bbb-b87a-4e2e-bc55-940edda975cc-kube-api-access-54klh\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.309950 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5508bbb-b87a-4e2e-bc55-940edda975cc-config-volume\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.311053 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5508bbb-b87a-4e2e-bc55-940edda975cc-config-volume\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.328266 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5508bbb-b87a-4e2e-bc55-940edda975cc-secret-volume\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.330482 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54klh\" (UniqueName: \"kubernetes.io/projected/a5508bbb-b87a-4e2e-bc55-940edda975cc-kube-api-access-54klh\") pod \"collect-profiles-29320710-9r4wd\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.460370 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:00 crc kubenswrapper[4799]: I0930 14:30:00.689409 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd"] Sep 30 14:30:01 crc kubenswrapper[4799]: I0930 14:30:01.687065 4799 generic.go:334] "Generic (PLEG): container finished" podID="a5508bbb-b87a-4e2e-bc55-940edda975cc" containerID="414693d8ed85d684c76c8a5bd0bc96d3fd2d1dbb8df38e5f1ad928d3dd2993fa" exitCode=0 Sep 30 14:30:01 crc kubenswrapper[4799]: I0930 14:30:01.687230 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" event={"ID":"a5508bbb-b87a-4e2e-bc55-940edda975cc","Type":"ContainerDied","Data":"414693d8ed85d684c76c8a5bd0bc96d3fd2d1dbb8df38e5f1ad928d3dd2993fa"} Sep 30 14:30:01 crc kubenswrapper[4799]: I0930 14:30:01.687334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" event={"ID":"a5508bbb-b87a-4e2e-bc55-940edda975cc","Type":"ContainerStarted","Data":"6d6017c7392c9d6eda648b549dd2107828795b23b0368b557fd5737785f71e36"} Sep 30 14:30:02 crc kubenswrapper[4799]: I0930 14:30:02.900993 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:30:02 crc kubenswrapper[4799]: I0930 14:30:02.948309 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5508bbb-b87a-4e2e-bc55-940edda975cc-secret-volume\") pod \"a5508bbb-b87a-4e2e-bc55-940edda975cc\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " Sep 30 14:30:02 crc kubenswrapper[4799]: I0930 14:30:02.948376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54klh\" (UniqueName: \"kubernetes.io/projected/a5508bbb-b87a-4e2e-bc55-940edda975cc-kube-api-access-54klh\") pod \"a5508bbb-b87a-4e2e-bc55-940edda975cc\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " Sep 30 14:30:02 crc kubenswrapper[4799]: I0930 14:30:02.948407 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5508bbb-b87a-4e2e-bc55-940edda975cc-config-volume\") pod \"a5508bbb-b87a-4e2e-bc55-940edda975cc\" (UID: \"a5508bbb-b87a-4e2e-bc55-940edda975cc\") " Sep 30 14:30:02 crc kubenswrapper[4799]: I0930 14:30:02.949320 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5508bbb-b87a-4e2e-bc55-940edda975cc-config-volume" (OuterVolumeSpecName: "config-volume") pod "a5508bbb-b87a-4e2e-bc55-940edda975cc" (UID: "a5508bbb-b87a-4e2e-bc55-940edda975cc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:30:02 crc kubenswrapper[4799]: I0930 14:30:02.954510 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5508bbb-b87a-4e2e-bc55-940edda975cc-kube-api-access-54klh" (OuterVolumeSpecName: "kube-api-access-54klh") pod "a5508bbb-b87a-4e2e-bc55-940edda975cc" (UID: "a5508bbb-b87a-4e2e-bc55-940edda975cc"). InnerVolumeSpecName "kube-api-access-54klh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:30:02 crc kubenswrapper[4799]: I0930 14:30:02.954755 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5508bbb-b87a-4e2e-bc55-940edda975cc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a5508bbb-b87a-4e2e-bc55-940edda975cc" (UID: "a5508bbb-b87a-4e2e-bc55-940edda975cc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:30:03 crc kubenswrapper[4799]: I0930 14:30:03.049551 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5508bbb-b87a-4e2e-bc55-940edda975cc-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:03 crc kubenswrapper[4799]: I0930 14:30:03.049600 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54klh\" (UniqueName: \"kubernetes.io/projected/a5508bbb-b87a-4e2e-bc55-940edda975cc-kube-api-access-54klh\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:03 crc kubenswrapper[4799]: I0930 14:30:03.049613 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5508bbb-b87a-4e2e-bc55-940edda975cc-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:30:03 crc kubenswrapper[4799]: I0930 14:30:03.698470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" event={"ID":"a5508bbb-b87a-4e2e-bc55-940edda975cc","Type":"ContainerDied","Data":"6d6017c7392c9d6eda648b549dd2107828795b23b0368b557fd5737785f71e36"} Sep 30 14:30:03 crc kubenswrapper[4799]: I0930 14:30:03.698854 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d6017c7392c9d6eda648b549dd2107828795b23b0368b557fd5737785f71e36" Sep 30 14:30:03 crc kubenswrapper[4799]: I0930 14:30:03.698913 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.648884 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.649504 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.859564 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vdjtl"] Sep 30 14:31:29 crc kubenswrapper[4799]: E0930 14:31:29.859825 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5508bbb-b87a-4e2e-bc55-940edda975cc" containerName="collect-profiles" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.859843 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5508bbb-b87a-4e2e-bc55-940edda975cc" containerName="collect-profiles" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.859969 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5508bbb-b87a-4e2e-bc55-940edda975cc" containerName="collect-profiles" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.860431 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-vdjtl" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.862980 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-h2x9z" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.865240 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.865323 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.877182 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vdjtl"] Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.886760 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mn5dh"] Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.887759 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-mn5dh" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.890418 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-h5ckn" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.896450 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqb56\" (UniqueName: \"kubernetes.io/projected/9146b875-e4c0-451d-8991-858332cc39d7-kube-api-access-pqb56\") pod \"cert-manager-cainjector-7f985d654d-vdjtl\" (UID: \"9146b875-e4c0-451d-8991-858332cc39d7\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vdjtl" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.896525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7g2b\" (UniqueName: \"kubernetes.io/projected/e3b68d27-f024-4124-99d3-d16f2171c911-kube-api-access-f7g2b\") pod \"cert-manager-5b446d88c5-mn5dh\" (UID: \"e3b68d27-f024-4124-99d3-d16f2171c911\") " pod="cert-manager/cert-manager-5b446d88c5-mn5dh" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.927053 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-94zcs"] Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.928637 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.931870 4799 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hvwlg" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.940840 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-94zcs"] Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.953037 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mn5dh"] Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.997254 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7g2b\" (UniqueName: \"kubernetes.io/projected/e3b68d27-f024-4124-99d3-d16f2171c911-kube-api-access-f7g2b\") pod \"cert-manager-5b446d88c5-mn5dh\" (UID: \"e3b68d27-f024-4124-99d3-d16f2171c911\") " pod="cert-manager/cert-manager-5b446d88c5-mn5dh" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.997318 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nksmt\" (UniqueName: \"kubernetes.io/projected/8bb043b3-1966-4b5e-bc13-476e14eb927e-kube-api-access-nksmt\") pod \"cert-manager-webhook-5655c58dd6-94zcs\" (UID: \"8bb043b3-1966-4b5e-bc13-476e14eb927e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" Sep 30 14:31:29 crc kubenswrapper[4799]: I0930 14:31:29.997396 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqb56\" (UniqueName: \"kubernetes.io/projected/9146b875-e4c0-451d-8991-858332cc39d7-kube-api-access-pqb56\") pod \"cert-manager-cainjector-7f985d654d-vdjtl\" (UID: \"9146b875-e4c0-451d-8991-858332cc39d7\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vdjtl" Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.016340 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7g2b\" (UniqueName: \"kubernetes.io/projected/e3b68d27-f024-4124-99d3-d16f2171c911-kube-api-access-f7g2b\") pod \"cert-manager-5b446d88c5-mn5dh\" (UID: \"e3b68d27-f024-4124-99d3-d16f2171c911\") " pod="cert-manager/cert-manager-5b446d88c5-mn5dh" Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.016556 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqb56\" (UniqueName: \"kubernetes.io/projected/9146b875-e4c0-451d-8991-858332cc39d7-kube-api-access-pqb56\") pod \"cert-manager-cainjector-7f985d654d-vdjtl\" (UID: \"9146b875-e4c0-451d-8991-858332cc39d7\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vdjtl" Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.098147 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nksmt\" (UniqueName: \"kubernetes.io/projected/8bb043b3-1966-4b5e-bc13-476e14eb927e-kube-api-access-nksmt\") pod \"cert-manager-webhook-5655c58dd6-94zcs\" (UID: \"8bb043b3-1966-4b5e-bc13-476e14eb927e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.116559 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nksmt\" (UniqueName: \"kubernetes.io/projected/8bb043b3-1966-4b5e-bc13-476e14eb927e-kube-api-access-nksmt\") pod \"cert-manager-webhook-5655c58dd6-94zcs\" (UID: \"8bb043b3-1966-4b5e-bc13-476e14eb927e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.181942 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-vdjtl" Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.211081 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-mn5dh" Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.245882 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.439271 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vdjtl"] Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.452903 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.528982 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-mn5dh"] Sep 30 14:31:30 crc kubenswrapper[4799]: I0930 14:31:30.561298 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-94zcs"] Sep 30 14:31:30 crc kubenswrapper[4799]: W0930 14:31:30.566707 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bb043b3_1966_4b5e_bc13_476e14eb927e.slice/crio-97a3f98bb820ec97f952d1f6b2a8be8b887e3f9990e336a12c490bfc8aff5c00 WatchSource:0}: Error finding container 97a3f98bb820ec97f952d1f6b2a8be8b887e3f9990e336a12c490bfc8aff5c00: Status 404 returned error can't find the container with id 97a3f98bb820ec97f952d1f6b2a8be8b887e3f9990e336a12c490bfc8aff5c00 Sep 30 14:31:31 crc kubenswrapper[4799]: I0930 14:31:31.142577 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-vdjtl" event={"ID":"9146b875-e4c0-451d-8991-858332cc39d7","Type":"ContainerStarted","Data":"387e5526cc49f3d631e109148937fbf55db93c6f0c3dc5ea105b93334adfac02"} Sep 30 14:31:31 crc kubenswrapper[4799]: I0930 14:31:31.144037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" event={"ID":"8bb043b3-1966-4b5e-bc13-476e14eb927e","Type":"ContainerStarted","Data":"97a3f98bb820ec97f952d1f6b2a8be8b887e3f9990e336a12c490bfc8aff5c00"} Sep 30 14:31:31 crc kubenswrapper[4799]: I0930 14:31:31.145506 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-mn5dh" event={"ID":"e3b68d27-f024-4124-99d3-d16f2171c911","Type":"ContainerStarted","Data":"5008ad97da8cc4b0acd8a43513aa38b5e7e4a8a90584f1d949b92026c71a0b32"} Sep 30 14:31:34 crc kubenswrapper[4799]: I0930 14:31:34.171506 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-mn5dh" event={"ID":"e3b68d27-f024-4124-99d3-d16f2171c911","Type":"ContainerStarted","Data":"e7505211f22d7458a7b999821595bbca66e99389b0e8268637078703276bf819"} Sep 30 14:31:34 crc kubenswrapper[4799]: I0930 14:31:34.173780 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-vdjtl" event={"ID":"9146b875-e4c0-451d-8991-858332cc39d7","Type":"ContainerStarted","Data":"68f05220640c455d696911165c38f056b2adea61b0af5e0f5555fe9c8b652475"} Sep 30 14:31:34 crc kubenswrapper[4799]: I0930 14:31:34.175449 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" event={"ID":"8bb043b3-1966-4b5e-bc13-476e14eb927e","Type":"ContainerStarted","Data":"175f3698e1c05ec1bd857b64e2e9f48b596d8a617091cee67b5356366a99a905"} Sep 30 14:31:34 crc kubenswrapper[4799]: I0930 14:31:34.175589 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" Sep 30 14:31:34 crc kubenswrapper[4799]: I0930 14:31:34.192118 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-mn5dh" podStartSLOduration=1.89646764 podStartE2EDuration="5.192098412s" podCreationTimestamp="2025-09-30 14:31:29 +0000 UTC" firstStartedPulling="2025-09-30 14:31:30.546403465 +0000 UTC m=+712.630003892" lastFinishedPulling="2025-09-30 14:31:33.842034237 +0000 UTC m=+715.925634664" observedRunningTime="2025-09-30 14:31:34.189627029 +0000 UTC m=+716.273227466" watchObservedRunningTime="2025-09-30 14:31:34.192098412 +0000 UTC m=+716.275698839" Sep 30 14:31:34 crc kubenswrapper[4799]: I0930 14:31:34.219960 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-vdjtl" podStartSLOduration=1.838363165 podStartE2EDuration="5.219938727s" podCreationTimestamp="2025-09-30 14:31:29 +0000 UTC" firstStartedPulling="2025-09-30 14:31:30.452578174 +0000 UTC m=+712.536178601" lastFinishedPulling="2025-09-30 14:31:33.834153736 +0000 UTC m=+715.917754163" observedRunningTime="2025-09-30 14:31:34.216996211 +0000 UTC m=+716.300596658" watchObservedRunningTime="2025-09-30 14:31:34.219938727 +0000 UTC m=+716.303539154" Sep 30 14:31:34 crc kubenswrapper[4799]: I0930 14:31:34.240871 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" podStartSLOduration=1.98195521 podStartE2EDuration="5.24083841s" podCreationTimestamp="2025-09-30 14:31:29 +0000 UTC" firstStartedPulling="2025-09-30 14:31:30.568990162 +0000 UTC m=+712.652590589" lastFinishedPulling="2025-09-30 14:31:33.827873362 +0000 UTC m=+715.911473789" observedRunningTime="2025-09-30 14:31:34.236559024 +0000 UTC m=+716.320159451" watchObservedRunningTime="2025-09-30 14:31:34.24083841 +0000 UTC m=+716.324438837" Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.829621 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6bhnh"] Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.830509 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovn-controller" containerID="cri-o://25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0" gracePeriod=30 Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.830996 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="northd" containerID="cri-o://91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3" gracePeriod=30 Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.831051 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528" gracePeriod=30 Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.831154 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kube-rbac-proxy-node" containerID="cri-o://17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2" gracePeriod=30 Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.831225 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovn-acl-logging" containerID="cri-o://b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5" gracePeriod=30 Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.831373 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="nbdb" containerID="cri-o://9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" gracePeriod=30 Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.831428 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="sbdb" containerID="cri-o://7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" gracePeriod=30 Sep 30 14:31:39 crc kubenswrapper[4799]: E0930 14:31:39.877930 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 30 14:31:39 crc kubenswrapper[4799]: E0930 14:31:39.885532 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 30 14:31:39 crc kubenswrapper[4799]: E0930 14:31:39.885556 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 30 14:31:39 crc kubenswrapper[4799]: E0930 14:31:39.892242 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 30 14:31:39 crc kubenswrapper[4799]: E0930 14:31:39.892927 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 30 14:31:39 crc kubenswrapper[4799]: E0930 14:31:39.892961 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="nbdb" Sep 30 14:31:39 crc kubenswrapper[4799]: E0930 14:31:39.894991 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 30 14:31:39 crc kubenswrapper[4799]: E0930 14:31:39.895049 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="sbdb" Sep 30 14:31:39 crc kubenswrapper[4799]: I0930 14:31:39.906089 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" containerID="cri-o://c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d" gracePeriod=30 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.208504 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/2.log" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.209084 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/1.log" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.209131 4799 generic.go:334] "Generic (PLEG): container finished" podID="5517db4f-9237-4b92-a672-6fa7afcd7e3f" containerID="52713cf3a77276c7a876e200beda5ecf6f3f1a5e784f0f9c2bb82d70e6d3f2b9" exitCode=2 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.209194 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sjjm8" event={"ID":"5517db4f-9237-4b92-a672-6fa7afcd7e3f","Type":"ContainerDied","Data":"52713cf3a77276c7a876e200beda5ecf6f3f1a5e784f0f9c2bb82d70e6d3f2b9"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.209245 4799 scope.go:117] "RemoveContainer" containerID="f6c8523bf8c18bcf854a04500f5f4b9dee82a6437885d44914b973343b4f90bb" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.209852 4799 scope.go:117] "RemoveContainer" containerID="52713cf3a77276c7a876e200beda5ecf6f3f1a5e784f0f9c2bb82d70e6d3f2b9" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.210047 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-sjjm8_openshift-multus(5517db4f-9237-4b92-a672-6fa7afcd7e3f)\"" pod="openshift-multus/multus-sjjm8" podUID="5517db4f-9237-4b92-a672-6fa7afcd7e3f" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.212051 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/3.log" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.214425 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovnkube-controller/3.log" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.215563 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovn-acl-logging/0.log" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.216160 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovn-controller/0.log" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.220224 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.220819 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovn-acl-logging/0.log" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.221339 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6bhnh_fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/ovn-controller/0.log" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225044 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d" exitCode=0 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225099 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" exitCode=0 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225110 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" exitCode=0 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225122 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3" exitCode=0 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225131 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528" exitCode=0 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225140 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2" exitCode=0 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225149 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5" exitCode=143 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225160 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerID="25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0" exitCode=143 Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225141 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225216 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225240 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225257 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225275 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225306 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225321 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225328 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225335 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225341 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225349 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225356 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225363 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225369 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225375 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225384 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225395 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225402 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225408 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225415 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225422 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225429 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225436 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225443 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225450 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225458 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225472 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225485 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225495 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225504 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225512 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225521 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225530 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225540 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225548 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225556 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225565 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225578 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" event={"ID":"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b","Type":"ContainerDied","Data":"944a76bbfb120710fbc9a2e94267336fa84172bc5b80c5b2eababbf3442fbd74"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225593 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225605 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225613 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225622 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225629 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225636 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225667 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225675 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225682 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.225688 4799 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.243640 4799 scope.go:117] "RemoveContainer" containerID="c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.254394 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-94zcs" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.268021 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.289176 4799 scope.go:117] "RemoveContainer" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.308089 4799 scope.go:117] "RemoveContainer" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.324974 4799 scope.go:117] "RemoveContainer" containerID="91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.332986 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovn-node-metrics-cert\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333038 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-netd\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333068 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-ovn\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333095 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333145 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-script-lib\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333178 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-slash\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333220 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-systemd-units\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333243 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-node-log\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333291 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-bin\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333325 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-etc-openvswitch\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-systemd\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333354 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-slash" (OuterVolumeSpecName: "host-slash") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333395 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-ovn-kubernetes\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333420 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333441 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-log-socket\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333462 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333483 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-openvswitch\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333495 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333508 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-var-lib-openvswitch\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333520 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-node-log" (OuterVolumeSpecName: "node-log") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333531 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-netns\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333551 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333558 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-kubelet\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333580 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333618 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-config\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333663 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-env-overrides\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333703 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh9hh\" (UniqueName: \"kubernetes.io/projected/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-kube-api-access-gh9hh\") pod \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\" (UID: \"fcc6c6df-d8b4-4d81-9aa5-8bc35533304b\") " Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.333823 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334040 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334379 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334419 4799 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-netns\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334441 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-log-socket" (OuterVolumeSpecName: "log-socket") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334446 4799 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334465 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334480 4799 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-slash\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334491 4799 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-systemd-units\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334501 4799 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-node-log\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334513 4799 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334523 4799 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334534 4799 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334878 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334912 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.334966 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.335040 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.335111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.335510 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.341085 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tcllm"] Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.341964 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-kube-api-access-gh9hh" (OuterVolumeSpecName: "kube-api-access-gh9hh") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "kube-api-access-gh9hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.342146 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.342707 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovn-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.342836 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovn-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.342916 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="nbdb" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.342987 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="nbdb" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.343077 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kube-rbac-proxy-node" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.343148 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kube-rbac-proxy-node" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.343220 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovn-acl-logging" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.343294 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovn-acl-logging" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.343364 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.343436 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.343512 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.343593 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.343697 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kubecfg-setup" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.343785 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kubecfg-setup" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.343856 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.343918 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.344043 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="northd" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344126 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="northd" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.344209 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="sbdb" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344301 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="sbdb" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.344362 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344415 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344582 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344698 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344779 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344873 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="northd" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344940 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="nbdb" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.344996 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovn-acl-logging" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.345051 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovn-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.345140 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="sbdb" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.345204 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kube-rbac-proxy-ovn-metrics" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.345259 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="kube-rbac-proxy-node" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.345413 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.345472 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.345528 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.345590 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.345782 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.345851 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" containerName="ovnkube-controller" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.346406 4799 scope.go:117] "RemoveContainer" containerID="ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.348479 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.351147 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" (UID: "fcc6c6df-d8b4-4d81-9aa5-8bc35533304b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.366549 4799 scope.go:117] "RemoveContainer" containerID="17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.382995 4799 scope.go:117] "RemoveContainer" containerID="b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.396180 4799 scope.go:117] "RemoveContainer" containerID="25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.422008 4799 scope.go:117] "RemoveContainer" containerID="aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436156 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-run-ovn-kubernetes\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436232 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-env-overrides\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436298 4799 scope.go:117] "RemoveContainer" containerID="c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436319 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-etc-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436475 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-slash\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436528 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovn-node-metrics-cert\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436552 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovnkube-config\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436577 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovnkube-script-lib\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436621 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-kubelet\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436754 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-systemd\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436808 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-ovn\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436844 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-cni-bin\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436875 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436909 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-node-log\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436941 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.436988 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-systemd-units\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437009 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-var-lib-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437030 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-cni-netd\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437053 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9mgg\" (UniqueName: \"kubernetes.io/projected/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-kube-api-access-q9mgg\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437080 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-run-netns\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437097 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-log-socket\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437165 4799 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-log-socket\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437182 4799 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437198 4799 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437210 4799 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-kubelet\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437223 4799 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437235 4799 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437246 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh9hh\" (UniqueName: \"kubernetes.io/projected/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-kube-api-access-gh9hh\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437257 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437269 4799 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437280 4799 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.437294 4799 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b-run-systemd\") on node \"crc\" DevicePath \"\"" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.438963 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": container with ID starting with c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d not found: ID does not exist" containerID="c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.439023 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} err="failed to get container status \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": rpc error: code = NotFound desc = could not find container \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": container with ID starting with c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.439054 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.439742 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": container with ID starting with 7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4 not found: ID does not exist" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.439791 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} err="failed to get container status \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": rpc error: code = NotFound desc = could not find container \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": container with ID starting with 7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.439834 4799 scope.go:117] "RemoveContainer" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.440157 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": container with ID starting with 7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde not found: ID does not exist" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.440180 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} err="failed to get container status \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": rpc error: code = NotFound desc = could not find container \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": container with ID starting with 7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.440195 4799 scope.go:117] "RemoveContainer" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.440515 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": container with ID starting with 9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc not found: ID does not exist" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.440556 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} err="failed to get container status \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": rpc error: code = NotFound desc = could not find container \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": container with ID starting with 9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.440577 4799 scope.go:117] "RemoveContainer" containerID="91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.441354 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": container with ID starting with 91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3 not found: ID does not exist" containerID="91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.441379 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} err="failed to get container status \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": rpc error: code = NotFound desc = could not find container \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": container with ID starting with 91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.441402 4799 scope.go:117] "RemoveContainer" containerID="ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.441695 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": container with ID starting with ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528 not found: ID does not exist" containerID="ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.441724 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} err="failed to get container status \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": rpc error: code = NotFound desc = could not find container \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": container with ID starting with ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.441742 4799 scope.go:117] "RemoveContainer" containerID="17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.441988 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": container with ID starting with 17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2 not found: ID does not exist" containerID="17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.442016 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} err="failed to get container status \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": rpc error: code = NotFound desc = could not find container \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": container with ID starting with 17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.442036 4799 scope.go:117] "RemoveContainer" containerID="b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.442284 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": container with ID starting with b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5 not found: ID does not exist" containerID="b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.442313 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} err="failed to get container status \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": rpc error: code = NotFound desc = could not find container \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": container with ID starting with b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.442331 4799 scope.go:117] "RemoveContainer" containerID="25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.442591 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": container with ID starting with 25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0 not found: ID does not exist" containerID="25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.442640 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} err="failed to get container status \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": rpc error: code = NotFound desc = could not find container \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": container with ID starting with 25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.442685 4799 scope.go:117] "RemoveContainer" containerID="aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff" Sep 30 14:31:40 crc kubenswrapper[4799]: E0930 14:31:40.443193 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": container with ID starting with aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff not found: ID does not exist" containerID="aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.443231 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} err="failed to get container status \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": rpc error: code = NotFound desc = could not find container \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": container with ID starting with aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.443269 4799 scope.go:117] "RemoveContainer" containerID="c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.444032 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} err="failed to get container status \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": rpc error: code = NotFound desc = could not find container \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": container with ID starting with c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.444056 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.444309 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} err="failed to get container status \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": rpc error: code = NotFound desc = could not find container \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": container with ID starting with 7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.444336 4799 scope.go:117] "RemoveContainer" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.444692 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} err="failed to get container status \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": rpc error: code = NotFound desc = could not find container \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": container with ID starting with 7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.444714 4799 scope.go:117] "RemoveContainer" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.444975 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} err="failed to get container status \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": rpc error: code = NotFound desc = could not find container \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": container with ID starting with 9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.445004 4799 scope.go:117] "RemoveContainer" containerID="91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.445302 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} err="failed to get container status \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": rpc error: code = NotFound desc = could not find container \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": container with ID starting with 91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.445326 4799 scope.go:117] "RemoveContainer" containerID="ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.445580 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} err="failed to get container status \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": rpc error: code = NotFound desc = could not find container \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": container with ID starting with ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.445605 4799 scope.go:117] "RemoveContainer" containerID="17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.445912 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} err="failed to get container status \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": rpc error: code = NotFound desc = could not find container \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": container with ID starting with 17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.445932 4799 scope.go:117] "RemoveContainer" containerID="b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.446177 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} err="failed to get container status \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": rpc error: code = NotFound desc = could not find container \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": container with ID starting with b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.446231 4799 scope.go:117] "RemoveContainer" containerID="25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.446643 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} err="failed to get container status \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": rpc error: code = NotFound desc = could not find container \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": container with ID starting with 25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.446702 4799 scope.go:117] "RemoveContainer" containerID="aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.447082 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} err="failed to get container status \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": rpc error: code = NotFound desc = could not find container \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": container with ID starting with aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.447102 4799 scope.go:117] "RemoveContainer" containerID="c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.447535 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} err="failed to get container status \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": rpc error: code = NotFound desc = could not find container \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": container with ID starting with c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.447555 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.447851 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} err="failed to get container status \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": rpc error: code = NotFound desc = could not find container \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": container with ID starting with 7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.447879 4799 scope.go:117] "RemoveContainer" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.448289 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} err="failed to get container status \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": rpc error: code = NotFound desc = could not find container \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": container with ID starting with 7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.448311 4799 scope.go:117] "RemoveContainer" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.448550 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} err="failed to get container status \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": rpc error: code = NotFound desc = could not find container \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": container with ID starting with 9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.448578 4799 scope.go:117] "RemoveContainer" containerID="91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.448861 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} err="failed to get container status \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": rpc error: code = NotFound desc = could not find container \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": container with ID starting with 91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.448890 4799 scope.go:117] "RemoveContainer" containerID="ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.449148 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} err="failed to get container status \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": rpc error: code = NotFound desc = could not find container \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": container with ID starting with ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.449182 4799 scope.go:117] "RemoveContainer" containerID="17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.449494 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} err="failed to get container status \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": rpc error: code = NotFound desc = could not find container \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": container with ID starting with 17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.449523 4799 scope.go:117] "RemoveContainer" containerID="b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.450134 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} err="failed to get container status \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": rpc error: code = NotFound desc = could not find container \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": container with ID starting with b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.450158 4799 scope.go:117] "RemoveContainer" containerID="25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.450490 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} err="failed to get container status \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": rpc error: code = NotFound desc = could not find container \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": container with ID starting with 25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.450516 4799 scope.go:117] "RemoveContainer" containerID="aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.450869 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} err="failed to get container status \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": rpc error: code = NotFound desc = could not find container \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": container with ID starting with aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.450898 4799 scope.go:117] "RemoveContainer" containerID="c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.451316 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d"} err="failed to get container status \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": rpc error: code = NotFound desc = could not find container \"c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d\": container with ID starting with c8f2cb3aff2562e8d15068180f7d9d785f17082515be14d1e2da04f50bd7264d not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.451337 4799 scope.go:117] "RemoveContainer" containerID="7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.451681 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4"} err="failed to get container status \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": rpc error: code = NotFound desc = could not find container \"7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4\": container with ID starting with 7e567c653e032cc0d54e2dee8889d2aa44a6363a185765b465aef00b17ec97b4 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.451713 4799 scope.go:117] "RemoveContainer" containerID="7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.452107 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde"} err="failed to get container status \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": rpc error: code = NotFound desc = could not find container \"7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde\": container with ID starting with 7b7b7432be06377e24845d03bbe93eaade85b537eeb0b193a2cf3867de891fde not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.452137 4799 scope.go:117] "RemoveContainer" containerID="9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.452418 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc"} err="failed to get container status \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": rpc error: code = NotFound desc = could not find container \"9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc\": container with ID starting with 9aa6616b917d571687ffd5d8b447b371d595cf978b3461f8233492c8d943ffcc not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.452446 4799 scope.go:117] "RemoveContainer" containerID="91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.452745 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3"} err="failed to get container status \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": rpc error: code = NotFound desc = could not find container \"91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3\": container with ID starting with 91e33d920afa94a9ab9b41f77ecde5da19f7e2f4b8375c567ca194d25b24d4b3 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.452792 4799 scope.go:117] "RemoveContainer" containerID="ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.453026 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528"} err="failed to get container status \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": rpc error: code = NotFound desc = could not find container \"ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528\": container with ID starting with ea86e9aead1fa009e4698247a9dd38b78aef766df1f6a9d9cd353022ec84f528 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.453053 4799 scope.go:117] "RemoveContainer" containerID="17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.453347 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2"} err="failed to get container status \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": rpc error: code = NotFound desc = could not find container \"17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2\": container with ID starting with 17039261e16a95729f7cc11cb8f725f35eff251a95b0a752173f040ea46a77e2 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.453367 4799 scope.go:117] "RemoveContainer" containerID="b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.455158 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5"} err="failed to get container status \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": rpc error: code = NotFound desc = could not find container \"b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5\": container with ID starting with b655a2946e2cc7c4173c76c19d1e9715b08e51701634680289f4ac0c49f708f5 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.455221 4799 scope.go:117] "RemoveContainer" containerID="25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.460925 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0"} err="failed to get container status \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": rpc error: code = NotFound desc = could not find container \"25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0\": container with ID starting with 25e1085a24cba32bee3599e06cf55ef72bc39ebe31438741bc674715589f9cc0 not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.460976 4799 scope.go:117] "RemoveContainer" containerID="aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.461894 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff"} err="failed to get container status \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": rpc error: code = NotFound desc = could not find container \"aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff\": container with ID starting with aa5fe83c9f8585d35c93b3bb1b5f48425316ef5bb725d12630edffe8e3defeff not found: ID does not exist" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538558 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-etc-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538615 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-slash\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538673 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovn-node-metrics-cert\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538701 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovnkube-config\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538725 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovnkube-script-lib\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538722 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-etc-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538749 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-kubelet\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538843 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-systemd\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538876 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-ovn\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538899 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-ovn\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538836 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-slash\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538843 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-kubelet\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-cni-bin\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539034 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539084 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-node-log\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539117 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539170 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-systemd-units\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539194 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-var-lib-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539223 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-cni-netd\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539248 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9mgg\" (UniqueName: \"kubernetes.io/projected/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-kube-api-access-q9mgg\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-run-netns\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539304 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-log-socket\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-run-ovn-kubernetes\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539398 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-env-overrides\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539568 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovnkube-script-lib\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539613 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-systemd-units\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538873 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-systemd\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539675 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.538926 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-cni-bin\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539715 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-node-log\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539741 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-run-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539831 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-var-lib-openvswitch\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539884 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-log-socket\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539923 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-run-netns\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539947 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-env-overrides\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539954 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-cni-netd\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.539989 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-host-run-ovn-kubernetes\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.540686 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovnkube-config\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.544022 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-ovn-node-metrics-cert\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.557310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9mgg\" (UniqueName: \"kubernetes.io/projected/ac82e6c0-06ba-4161-a659-18a8b07b7dd6-kube-api-access-q9mgg\") pod \"ovnkube-node-tcllm\" (UID: \"ac82e6c0-06ba-4161-a659-18a8b07b7dd6\") " pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: I0930 14:31:40.665336 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:40 crc kubenswrapper[4799]: W0930 14:31:40.686088 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac82e6c0_06ba_4161_a659_18a8b07b7dd6.slice/crio-11bc40459f204e1c6028851bb4806e9bd3f0d672919f5d4ff9b88bd837f7b273 WatchSource:0}: Error finding container 11bc40459f204e1c6028851bb4806e9bd3f0d672919f5d4ff9b88bd837f7b273: Status 404 returned error can't find the container with id 11bc40459f204e1c6028851bb4806e9bd3f0d672919f5d4ff9b88bd837f7b273 Sep 30 14:31:41 crc kubenswrapper[4799]: I0930 14:31:41.232692 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/2.log" Sep 30 14:31:41 crc kubenswrapper[4799]: I0930 14:31:41.234865 4799 generic.go:334] "Generic (PLEG): container finished" podID="ac82e6c0-06ba-4161-a659-18a8b07b7dd6" containerID="328ab639e36c0d5b1b28779b2dff788e8f459eca97ca21ab34eabe183142d976" exitCode=0 Sep 30 14:31:41 crc kubenswrapper[4799]: I0930 14:31:41.234940 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerDied","Data":"328ab639e36c0d5b1b28779b2dff788e8f459eca97ca21ab34eabe183142d976"} Sep 30 14:31:41 crc kubenswrapper[4799]: I0930 14:31:41.234968 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"11bc40459f204e1c6028851bb4806e9bd3f0d672919f5d4ff9b88bd837f7b273"} Sep 30 14:31:41 crc kubenswrapper[4799]: I0930 14:31:41.237640 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6bhnh" Sep 30 14:31:41 crc kubenswrapper[4799]: I0930 14:31:41.312850 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6bhnh"] Sep 30 14:31:41 crc kubenswrapper[4799]: I0930 14:31:41.329214 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6bhnh"] Sep 30 14:31:42 crc kubenswrapper[4799]: I0930 14:31:42.247796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"a91afd0853bbf0a3ee98740f9f3e0fd6b4b4d8ac55a2cbb2bf6e1d8225beb287"} Sep 30 14:31:42 crc kubenswrapper[4799]: I0930 14:31:42.248382 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"80c8fa48e6398ad35817cbcf355087fb2ba2e9b60b745f5888c6fdeed8cfffb4"} Sep 30 14:31:42 crc kubenswrapper[4799]: I0930 14:31:42.248398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"4badabfc6b935e5e4dc603f698708228abf050e13ed69f2b8ee04c7fe7d73194"} Sep 30 14:31:42 crc kubenswrapper[4799]: I0930 14:31:42.248409 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"6053f1f2f772da871a9c40644a396c0b4b963305d91ab82d39b2322572006b91"} Sep 30 14:31:42 crc kubenswrapper[4799]: I0930 14:31:42.248418 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"ee3e50771667914d04a1111f07218e595a966c11d68cbf290e7170e287985274"} Sep 30 14:31:42 crc kubenswrapper[4799]: I0930 14:31:42.248426 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"86bb6bd903b661eebea498bbbd1191a9b844e724f3da948b41ea7486662fb904"} Sep 30 14:31:42 crc kubenswrapper[4799]: I0930 14:31:42.510625 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcc6c6df-d8b4-4d81-9aa5-8bc35533304b" path="/var/lib/kubelet/pods/fcc6c6df-d8b4-4d81-9aa5-8bc35533304b/volumes" Sep 30 14:31:44 crc kubenswrapper[4799]: I0930 14:31:44.264720 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"206e5da48f0039bb01c6d4211abf5a352ed30dbc100f7c37e63a3ea1cc977c50"} Sep 30 14:31:47 crc kubenswrapper[4799]: I0930 14:31:47.284917 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" event={"ID":"ac82e6c0-06ba-4161-a659-18a8b07b7dd6","Type":"ContainerStarted","Data":"82b212f7ed7d26975ef3ad8dbb695c1b07a010c707fd446487ec06084a06d007"} Sep 30 14:31:47 crc kubenswrapper[4799]: I0930 14:31:47.286537 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:47 crc kubenswrapper[4799]: I0930 14:31:47.286555 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:47 crc kubenswrapper[4799]: I0930 14:31:47.286564 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:47 crc kubenswrapper[4799]: I0930 14:31:47.311129 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:47 crc kubenswrapper[4799]: I0930 14:31:47.319576 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" podStartSLOduration=7.319558794 podStartE2EDuration="7.319558794s" podCreationTimestamp="2025-09-30 14:31:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:31:47.311848659 +0000 UTC m=+729.395449086" watchObservedRunningTime="2025-09-30 14:31:47.319558794 +0000 UTC m=+729.403159221" Sep 30 14:31:47 crc kubenswrapper[4799]: I0930 14:31:47.323700 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:31:53 crc kubenswrapper[4799]: I0930 14:31:53.503196 4799 scope.go:117] "RemoveContainer" containerID="52713cf3a77276c7a876e200beda5ecf6f3f1a5e784f0f9c2bb82d70e6d3f2b9" Sep 30 14:31:54 crc kubenswrapper[4799]: I0930 14:31:54.326071 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-sjjm8_5517db4f-9237-4b92-a672-6fa7afcd7e3f/kube-multus/2.log" Sep 30 14:31:54 crc kubenswrapper[4799]: I0930 14:31:54.326922 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-sjjm8" event={"ID":"5517db4f-9237-4b92-a672-6fa7afcd7e3f","Type":"ContainerStarted","Data":"835b434f567cb84a581f0cfe1b679ecd7ff094f878e8f59890544cb72a3b2cc1"} Sep 30 14:31:59 crc kubenswrapper[4799]: I0930 14:31:59.649303 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:31:59 crc kubenswrapper[4799]: I0930 14:31:59.650751 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:32:06 crc kubenswrapper[4799]: I0930 14:32:06.616234 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmw7h"] Sep 30 14:32:06 crc kubenswrapper[4799]: I0930 14:32:06.617001 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" podUID="a3981843-fd55-41a8-8377-a741fcd231fb" containerName="controller-manager" containerID="cri-o://0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969" gracePeriod=30 Sep 30 14:32:06 crc kubenswrapper[4799]: I0930 14:32:06.708072 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k"] Sep 30 14:32:06 crc kubenswrapper[4799]: I0930 14:32:06.708282 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" podUID="9fabab36-ee1e-4d92-b99c-b07ae87891be" containerName="route-controller-manager" containerID="cri-o://2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca" gracePeriod=30 Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.230246 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.243721 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.382707 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66k57\" (UniqueName: \"kubernetes.io/projected/a3981843-fd55-41a8-8377-a741fcd231fb-kube-api-access-66k57\") pod \"a3981843-fd55-41a8-8377-a741fcd231fb\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.382777 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqs2c\" (UniqueName: \"kubernetes.io/projected/9fabab36-ee1e-4d92-b99c-b07ae87891be-kube-api-access-lqs2c\") pod \"9fabab36-ee1e-4d92-b99c-b07ae87891be\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.382805 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-proxy-ca-bundles\") pod \"a3981843-fd55-41a8-8377-a741fcd231fb\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.382857 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-client-ca\") pod \"a3981843-fd55-41a8-8377-a741fcd231fb\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.382907 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3981843-fd55-41a8-8377-a741fcd231fb-serving-cert\") pod \"a3981843-fd55-41a8-8377-a741fcd231fb\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.382941 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-client-ca\") pod \"9fabab36-ee1e-4d92-b99c-b07ae87891be\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.382982 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-config\") pod \"a3981843-fd55-41a8-8377-a741fcd231fb\" (UID: \"a3981843-fd55-41a8-8377-a741fcd231fb\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.383013 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fabab36-ee1e-4d92-b99c-b07ae87891be-serving-cert\") pod \"9fabab36-ee1e-4d92-b99c-b07ae87891be\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.383044 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-config\") pod \"9fabab36-ee1e-4d92-b99c-b07ae87891be\" (UID: \"9fabab36-ee1e-4d92-b99c-b07ae87891be\") " Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.384335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a3981843-fd55-41a8-8377-a741fcd231fb" (UID: "a3981843-fd55-41a8-8377-a741fcd231fb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.384354 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-client-ca" (OuterVolumeSpecName: "client-ca") pod "a3981843-fd55-41a8-8377-a741fcd231fb" (UID: "a3981843-fd55-41a8-8377-a741fcd231fb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.384808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-config" (OuterVolumeSpecName: "config") pod "a3981843-fd55-41a8-8377-a741fcd231fb" (UID: "a3981843-fd55-41a8-8377-a741fcd231fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.386100 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-client-ca" (OuterVolumeSpecName: "client-ca") pod "9fabab36-ee1e-4d92-b99c-b07ae87891be" (UID: "9fabab36-ee1e-4d92-b99c-b07ae87891be"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.386196 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-config" (OuterVolumeSpecName: "config") pod "9fabab36-ee1e-4d92-b99c-b07ae87891be" (UID: "9fabab36-ee1e-4d92-b99c-b07ae87891be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.392466 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fabab36-ee1e-4d92-b99c-b07ae87891be-kube-api-access-lqs2c" (OuterVolumeSpecName: "kube-api-access-lqs2c") pod "9fabab36-ee1e-4d92-b99c-b07ae87891be" (UID: "9fabab36-ee1e-4d92-b99c-b07ae87891be"). InnerVolumeSpecName "kube-api-access-lqs2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.392578 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3981843-fd55-41a8-8377-a741fcd231fb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a3981843-fd55-41a8-8377-a741fcd231fb" (UID: "a3981843-fd55-41a8-8377-a741fcd231fb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.392611 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fabab36-ee1e-4d92-b99c-b07ae87891be-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9fabab36-ee1e-4d92-b99c-b07ae87891be" (UID: "9fabab36-ee1e-4d92-b99c-b07ae87891be"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.393191 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3981843-fd55-41a8-8377-a741fcd231fb-kube-api-access-66k57" (OuterVolumeSpecName: "kube-api-access-66k57") pod "a3981843-fd55-41a8-8377-a741fcd231fb" (UID: "a3981843-fd55-41a8-8377-a741fcd231fb"). InnerVolumeSpecName "kube-api-access-66k57". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.404237 4799 generic.go:334] "Generic (PLEG): container finished" podID="9fabab36-ee1e-4d92-b99c-b07ae87891be" containerID="2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca" exitCode=0 Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.404325 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" event={"ID":"9fabab36-ee1e-4d92-b99c-b07ae87891be","Type":"ContainerDied","Data":"2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca"} Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.404361 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" event={"ID":"9fabab36-ee1e-4d92-b99c-b07ae87891be","Type":"ContainerDied","Data":"0a9b4bcbc351a7a5c77d7f232fb662559f925515ba44f307563e9a0b6ec04324"} Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.404385 4799 scope.go:117] "RemoveContainer" containerID="2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.404535 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.410032 4799 generic.go:334] "Generic (PLEG): container finished" podID="a3981843-fd55-41a8-8377-a741fcd231fb" containerID="0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969" exitCode=0 Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.410076 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" event={"ID":"a3981843-fd55-41a8-8377-a741fcd231fb","Type":"ContainerDied","Data":"0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969"} Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.410124 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" event={"ID":"a3981843-fd55-41a8-8377-a741fcd231fb","Type":"ContainerDied","Data":"21544cd61b5b34562039e3b1fee260249bcd0faf29482405de90e86ac9d286f2"} Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.410092 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rmw7h" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.446300 4799 scope.go:117] "RemoveContainer" containerID="2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.450202 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmw7h"] Sep 30 14:32:07 crc kubenswrapper[4799]: E0930 14:32:07.451717 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca\": container with ID starting with 2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca not found: ID does not exist" containerID="2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.451775 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca"} err="failed to get container status \"2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca\": rpc error: code = NotFound desc = could not find container \"2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca\": container with ID starting with 2b8ff37921dc2210ceef0fec646a5e0daf6b1643eff0cad733566dd7aab4f0ca not found: ID does not exist" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.451813 4799 scope.go:117] "RemoveContainer" containerID="0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.455212 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rmw7h"] Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.469712 4799 scope.go:117] "RemoveContainer" containerID="0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969" Sep 30 14:32:07 crc kubenswrapper[4799]: E0930 14:32:07.477176 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969\": container with ID starting with 0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969 not found: ID does not exist" containerID="0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.477333 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969"} err="failed to get container status \"0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969\": rpc error: code = NotFound desc = could not find container \"0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969\": container with ID starting with 0a8d5d1c262c7079b0ef46de3f7ed67f4ee40719cc5743a35d076977abaa4969 not found: ID does not exist" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.479836 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k"] Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.485914 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xkv5k"] Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493108 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66k57\" (UniqueName: \"kubernetes.io/projected/a3981843-fd55-41a8-8377-a741fcd231fb-kube-api-access-66k57\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493157 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqs2c\" (UniqueName: \"kubernetes.io/projected/9fabab36-ee1e-4d92-b99c-b07ae87891be-kube-api-access-lqs2c\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493174 4799 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493194 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493208 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3981843-fd55-41a8-8377-a741fcd231fb-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493220 4799 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-client-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493268 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3981843-fd55-41a8-8377-a741fcd231fb-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493287 4799 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fabab36-ee1e-4d92-b99c-b07ae87891be-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:07 crc kubenswrapper[4799]: I0930 14:32:07.493300 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fabab36-ee1e-4d92-b99c-b07ae87891be-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.212258 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw"] Sep 30 14:32:08 crc kubenswrapper[4799]: E0930 14:32:08.213681 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fabab36-ee1e-4d92-b99c-b07ae87891be" containerName="route-controller-manager" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.213760 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fabab36-ee1e-4d92-b99c-b07ae87891be" containerName="route-controller-manager" Sep 30 14:32:08 crc kubenswrapper[4799]: E0930 14:32:08.213860 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3981843-fd55-41a8-8377-a741fcd231fb" containerName="controller-manager" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.213924 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3981843-fd55-41a8-8377-a741fcd231fb" containerName="controller-manager" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.214098 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3981843-fd55-41a8-8377-a741fcd231fb" containerName="controller-manager" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.214180 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fabab36-ee1e-4d92-b99c-b07ae87891be" containerName="route-controller-manager" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.214702 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.217461 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.217566 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.217589 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.217804 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.217836 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.221485 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.224316 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw"] Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.238219 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76f55db76-gv52k"] Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.239086 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.241412 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.241939 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.241959 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.242020 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.242529 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.243611 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.256746 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76f55db76-gv52k"] Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.261264 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.403977 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e055c753-cf84-4832-9b80-e8c2a2f31bd1-config\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.404284 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e055c753-cf84-4832-9b80-e8c2a2f31bd1-serving-cert\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.404323 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-client-ca\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.404343 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-proxy-ca-bundles\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.404364 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41157dae-e986-4e2b-b8eb-5985dc7199ee-serving-cert\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.404407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-config\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.404440 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl25x\" (UniqueName: \"kubernetes.io/projected/41157dae-e986-4e2b-b8eb-5985dc7199ee-kube-api-access-bl25x\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.404461 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp4w2\" (UniqueName: \"kubernetes.io/projected/e055c753-cf84-4832-9b80-e8c2a2f31bd1-kube-api-access-wp4w2\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.404571 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e055c753-cf84-4832-9b80-e8c2a2f31bd1-client-ca\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505333 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-config\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl25x\" (UniqueName: \"kubernetes.io/projected/41157dae-e986-4e2b-b8eb-5985dc7199ee-kube-api-access-bl25x\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505592 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp4w2\" (UniqueName: \"kubernetes.io/projected/e055c753-cf84-4832-9b80-e8c2a2f31bd1-kube-api-access-wp4w2\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505611 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e055c753-cf84-4832-9b80-e8c2a2f31bd1-client-ca\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505634 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e055c753-cf84-4832-9b80-e8c2a2f31bd1-config\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505668 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e055c753-cf84-4832-9b80-e8c2a2f31bd1-serving-cert\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505697 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-client-ca\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505720 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-proxy-ca-bundles\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.505747 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41157dae-e986-4e2b-b8eb-5985dc7199ee-serving-cert\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.507742 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-config\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.507974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-client-ca\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.508135 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41157dae-e986-4e2b-b8eb-5985dc7199ee-proxy-ca-bundles\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.508487 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e055c753-cf84-4832-9b80-e8c2a2f31bd1-config\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.508691 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e055c753-cf84-4832-9b80-e8c2a2f31bd1-client-ca\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.512261 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e055c753-cf84-4832-9b80-e8c2a2f31bd1-serving-cert\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.515962 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fabab36-ee1e-4d92-b99c-b07ae87891be" path="/var/lib/kubelet/pods/9fabab36-ee1e-4d92-b99c-b07ae87891be/volumes" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.516242 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41157dae-e986-4e2b-b8eb-5985dc7199ee-serving-cert\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.516635 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3981843-fd55-41a8-8377-a741fcd231fb" path="/var/lib/kubelet/pods/a3981843-fd55-41a8-8377-a741fcd231fb/volumes" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.523127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl25x\" (UniqueName: \"kubernetes.io/projected/41157dae-e986-4e2b-b8eb-5985dc7199ee-kube-api-access-bl25x\") pod \"controller-manager-76f55db76-gv52k\" (UID: \"41157dae-e986-4e2b-b8eb-5985dc7199ee\") " pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.530721 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp4w2\" (UniqueName: \"kubernetes.io/projected/e055c753-cf84-4832-9b80-e8c2a2f31bd1-kube-api-access-wp4w2\") pod \"route-controller-manager-5959949b57-qchvw\" (UID: \"e055c753-cf84-4832-9b80-e8c2a2f31bd1\") " pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.536462 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.556771 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.795541 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw"] Sep 30 14:32:08 crc kubenswrapper[4799]: W0930 14:32:08.801839 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode055c753_cf84_4832_9b80_e8c2a2f31bd1.slice/crio-c4800e759929bb13a5f1bd57935c44fc95975f3613741c2c97bc70aee88ca72e WatchSource:0}: Error finding container c4800e759929bb13a5f1bd57935c44fc95975f3613741c2c97bc70aee88ca72e: Status 404 returned error can't find the container with id c4800e759929bb13a5f1bd57935c44fc95975f3613741c2c97bc70aee88ca72e Sep 30 14:32:08 crc kubenswrapper[4799]: I0930 14:32:08.852059 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76f55db76-gv52k"] Sep 30 14:32:08 crc kubenswrapper[4799]: W0930 14:32:08.861814 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41157dae_e986_4e2b_b8eb_5985dc7199ee.slice/crio-cc18790b95085485edf5cbb6ddb03e49533a080bd48dabeb5a61f9d98896f018 WatchSource:0}: Error finding container cc18790b95085485edf5cbb6ddb03e49533a080bd48dabeb5a61f9d98896f018: Status 404 returned error can't find the container with id cc18790b95085485edf5cbb6ddb03e49533a080bd48dabeb5a61f9d98896f018 Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.426472 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" event={"ID":"e055c753-cf84-4832-9b80-e8c2a2f31bd1","Type":"ContainerStarted","Data":"33f74fad605845bfc32b6acacadd785bd26a3e5c5adf1f32887fbf3ac63494a7"} Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.427694 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" event={"ID":"e055c753-cf84-4832-9b80-e8c2a2f31bd1","Type":"ContainerStarted","Data":"c4800e759929bb13a5f1bd57935c44fc95975f3613741c2c97bc70aee88ca72e"} Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.429946 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.431689 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" event={"ID":"41157dae-e986-4e2b-b8eb-5985dc7199ee","Type":"ContainerStarted","Data":"0c04e411dec2e0b2c9483910038f3b47ed1ff3cd1fc870398fd0ebed3185218a"} Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.431721 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" event={"ID":"41157dae-e986-4e2b-b8eb-5985dc7199ee","Type":"ContainerStarted","Data":"cc18790b95085485edf5cbb6ddb03e49533a080bd48dabeb5a61f9d98896f018"} Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.432003 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.448378 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.450906 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" podStartSLOduration=1.450888907 podStartE2EDuration="1.450888907s" podCreationTimestamp="2025-09-30 14:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:32:09.449866587 +0000 UTC m=+751.533467024" watchObservedRunningTime="2025-09-30 14:32:09.450888907 +0000 UTC m=+751.534489334" Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.687034 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5959949b57-qchvw" Sep 30 14:32:09 crc kubenswrapper[4799]: I0930 14:32:09.703903 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76f55db76-gv52k" podStartSLOduration=3.703883829 podStartE2EDuration="3.703883829s" podCreationTimestamp="2025-09-30 14:32:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:32:09.47215912 +0000 UTC m=+751.555759547" watchObservedRunningTime="2025-09-30 14:32:09.703883829 +0000 UTC m=+751.787484256" Sep 30 14:32:10 crc kubenswrapper[4799]: I0930 14:32:10.695312 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tcllm" Sep 30 14:32:18 crc kubenswrapper[4799]: I0930 14:32:18.143937 4799 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.521124 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh"] Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.522632 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.525575 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.540419 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh"] Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.689888 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.690029 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kscl5\" (UniqueName: \"kubernetes.io/projected/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-kube-api-access-kscl5\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.690128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.791184 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.791531 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kscl5\" (UniqueName: \"kubernetes.io/projected/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-kube-api-access-kscl5\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.791718 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.792229 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.792426 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.811745 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kscl5\" (UniqueName: \"kubernetes.io/projected/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-kube-api-access-kscl5\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:23 crc kubenswrapper[4799]: I0930 14:32:23.847170 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:24 crc kubenswrapper[4799]: I0930 14:32:24.242640 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh"] Sep 30 14:32:24 crc kubenswrapper[4799]: W0930 14:32:24.252304 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bd6c752_45af_4df7_9c3f_dadeeb43bd0c.slice/crio-0398fb422a85c47b19e85a4f87428e9f8d38262a2b5a459fafb20c9b3d99f544 WatchSource:0}: Error finding container 0398fb422a85c47b19e85a4f87428e9f8d38262a2b5a459fafb20c9b3d99f544: Status 404 returned error can't find the container with id 0398fb422a85c47b19e85a4f87428e9f8d38262a2b5a459fafb20c9b3d99f544 Sep 30 14:32:24 crc kubenswrapper[4799]: I0930 14:32:24.514811 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" event={"ID":"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c","Type":"ContainerStarted","Data":"87a4a2256229a5c1fa334b86da20bd57a6728964e9e5ff9619071085292aac62"} Sep 30 14:32:24 crc kubenswrapper[4799]: I0930 14:32:24.515173 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" event={"ID":"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c","Type":"ContainerStarted","Data":"0398fb422a85c47b19e85a4f87428e9f8d38262a2b5a459fafb20c9b3d99f544"} Sep 30 14:32:25 crc kubenswrapper[4799]: I0930 14:32:25.522728 4799 generic.go:334] "Generic (PLEG): container finished" podID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerID="87a4a2256229a5c1fa334b86da20bd57a6728964e9e5ff9619071085292aac62" exitCode=0 Sep 30 14:32:25 crc kubenswrapper[4799]: I0930 14:32:25.522796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" event={"ID":"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c","Type":"ContainerDied","Data":"87a4a2256229a5c1fa334b86da20bd57a6728964e9e5ff9619071085292aac62"} Sep 30 14:32:25 crc kubenswrapper[4799]: I0930 14:32:25.821526 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mp8vs"] Sep 30 14:32:25 crc kubenswrapper[4799]: I0930 14:32:25.822968 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:25 crc kubenswrapper[4799]: I0930 14:32:25.854529 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mp8vs"] Sep 30 14:32:25 crc kubenswrapper[4799]: I0930 14:32:25.919855 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz6zp\" (UniqueName: \"kubernetes.io/projected/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-kube-api-access-dz6zp\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:25 crc kubenswrapper[4799]: I0930 14:32:25.920015 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-catalog-content\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:25 crc kubenswrapper[4799]: I0930 14:32:25.920236 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-utilities\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:26 crc kubenswrapper[4799]: I0930 14:32:26.021282 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-utilities\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:26 crc kubenswrapper[4799]: I0930 14:32:26.021350 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz6zp\" (UniqueName: \"kubernetes.io/projected/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-kube-api-access-dz6zp\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:26 crc kubenswrapper[4799]: I0930 14:32:26.021386 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-catalog-content\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:26 crc kubenswrapper[4799]: I0930 14:32:26.022009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-utilities\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:26 crc kubenswrapper[4799]: I0930 14:32:26.022035 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-catalog-content\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:26 crc kubenswrapper[4799]: I0930 14:32:26.053756 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz6zp\" (UniqueName: \"kubernetes.io/projected/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-kube-api-access-dz6zp\") pod \"redhat-operators-mp8vs\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:26 crc kubenswrapper[4799]: I0930 14:32:26.143224 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:26 crc kubenswrapper[4799]: I0930 14:32:26.591589 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mp8vs"] Sep 30 14:32:27 crc kubenswrapper[4799]: I0930 14:32:27.535853 4799 generic.go:334] "Generic (PLEG): container finished" podID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerID="24fc3c1fe4c78db17f31e4ee9f7df3c004542dbc90b3feb4b3acbec591867dd3" exitCode=0 Sep 30 14:32:27 crc kubenswrapper[4799]: I0930 14:32:27.535890 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" event={"ID":"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c","Type":"ContainerDied","Data":"24fc3c1fe4c78db17f31e4ee9f7df3c004542dbc90b3feb4b3acbec591867dd3"} Sep 30 14:32:27 crc kubenswrapper[4799]: I0930 14:32:27.537746 4799 generic.go:334] "Generic (PLEG): container finished" podID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerID="da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f" exitCode=0 Sep 30 14:32:27 crc kubenswrapper[4799]: I0930 14:32:27.537959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mp8vs" event={"ID":"2ed54eba-fa9c-4754-82c1-d697d9bc40b8","Type":"ContainerDied","Data":"da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f"} Sep 30 14:32:27 crc kubenswrapper[4799]: I0930 14:32:27.538059 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mp8vs" event={"ID":"2ed54eba-fa9c-4754-82c1-d697d9bc40b8","Type":"ContainerStarted","Data":"64fec58d9ee16c59f66fb5a1a3fba1da89a0e0352a57339c6aa676991f1577e4"} Sep 30 14:32:28 crc kubenswrapper[4799]: I0930 14:32:28.545445 4799 generic.go:334] "Generic (PLEG): container finished" podID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerID="91927686ee48d3d58315a206eeea11db2523666e5058c2019111d43b564a3cb8" exitCode=0 Sep 30 14:32:28 crc kubenswrapper[4799]: I0930 14:32:28.545538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" event={"ID":"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c","Type":"ContainerDied","Data":"91927686ee48d3d58315a206eeea11db2523666e5058c2019111d43b564a3cb8"} Sep 30 14:32:28 crc kubenswrapper[4799]: I0930 14:32:28.547557 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mp8vs" event={"ID":"2ed54eba-fa9c-4754-82c1-d697d9bc40b8","Type":"ContainerStarted","Data":"8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b"} Sep 30 14:32:29 crc kubenswrapper[4799]: I0930 14:32:29.653316 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:32:29 crc kubenswrapper[4799]: I0930 14:32:29.655701 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:32:29 crc kubenswrapper[4799]: I0930 14:32:29.655764 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:32:29 crc kubenswrapper[4799]: I0930 14:32:29.656424 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e25c11d4641ce442c772191497cb466d8f447d176e97bcb309720e08dd67a66"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:32:29 crc kubenswrapper[4799]: I0930 14:32:29.656471 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://1e25c11d4641ce442c772191497cb466d8f447d176e97bcb309720e08dd67a66" gracePeriod=600 Sep 30 14:32:29 crc kubenswrapper[4799]: I0930 14:32:29.923968 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.082524 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kscl5\" (UniqueName: \"kubernetes.io/projected/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-kube-api-access-kscl5\") pod \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.082612 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-bundle\") pod \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.082642 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-util\") pod \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\" (UID: \"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c\") " Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.083350 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-bundle" (OuterVolumeSpecName: "bundle") pod "8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" (UID: "8bd6c752-45af-4df7-9c3f-dadeeb43bd0c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.093945 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-kube-api-access-kscl5" (OuterVolumeSpecName: "kube-api-access-kscl5") pod "8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" (UID: "8bd6c752-45af-4df7-9c3f-dadeeb43bd0c"). InnerVolumeSpecName "kube-api-access-kscl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.184405 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kscl5\" (UniqueName: \"kubernetes.io/projected/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-kube-api-access-kscl5\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.184448 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.350911 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-util" (OuterVolumeSpecName: "util") pod "8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" (UID: "8bd6c752-45af-4df7-9c3f-dadeeb43bd0c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.387369 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bd6c752-45af-4df7-9c3f-dadeeb43bd0c-util\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.562433 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" event={"ID":"8bd6c752-45af-4df7-9c3f-dadeeb43bd0c","Type":"ContainerDied","Data":"0398fb422a85c47b19e85a4f87428e9f8d38262a2b5a459fafb20c9b3d99f544"} Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.562489 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0398fb422a85c47b19e85a4f87428e9f8d38262a2b5a459fafb20c9b3d99f544" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.562491 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.565076 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="1e25c11d4641ce442c772191497cb466d8f447d176e97bcb309720e08dd67a66" exitCode=0 Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.565143 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"1e25c11d4641ce442c772191497cb466d8f447d176e97bcb309720e08dd67a66"} Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.565174 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"7be6e12e797fed54fc7aab60eaeaf0c5516974a723cfc906b94db6e0a39463b8"} Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.565191 4799 scope.go:117] "RemoveContainer" containerID="d7967cc87bf65a71f9ed110a689be544a6e7e0ea0717c62f31a311bad53a8a0f" Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.567714 4799 generic.go:334] "Generic (PLEG): container finished" podID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerID="8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b" exitCode=0 Sep 30 14:32:30 crc kubenswrapper[4799]: I0930 14:32:30.567808 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mp8vs" event={"ID":"2ed54eba-fa9c-4754-82c1-d697d9bc40b8","Type":"ContainerDied","Data":"8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b"} Sep 30 14:32:31 crc kubenswrapper[4799]: I0930 14:32:31.578591 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mp8vs" event={"ID":"2ed54eba-fa9c-4754-82c1-d697d9bc40b8","Type":"ContainerStarted","Data":"7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2"} Sep 30 14:32:31 crc kubenswrapper[4799]: I0930 14:32:31.599742 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mp8vs" podStartSLOduration=3.116008445 podStartE2EDuration="6.599719564s" podCreationTimestamp="2025-09-30 14:32:25 +0000 UTC" firstStartedPulling="2025-09-30 14:32:27.539711032 +0000 UTC m=+769.623311459" lastFinishedPulling="2025-09-30 14:32:31.023422151 +0000 UTC m=+773.107022578" observedRunningTime="2025-09-30 14:32:31.596800028 +0000 UTC m=+773.680400485" watchObservedRunningTime="2025-09-30 14:32:31.599719564 +0000 UTC m=+773.683319991" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.089613 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx"] Sep 30 14:32:35 crc kubenswrapper[4799]: E0930 14:32:35.090152 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerName="util" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.090165 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerName="util" Sep 30 14:32:35 crc kubenswrapper[4799]: E0930 14:32:35.090179 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerName="extract" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.090185 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerName="extract" Sep 30 14:32:35 crc kubenswrapper[4799]: E0930 14:32:35.090204 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerName="pull" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.090209 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerName="pull" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.090299 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bd6c752-45af-4df7-9c3f-dadeeb43bd0c" containerName="extract" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.090746 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.093078 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.093719 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.093730 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-wd788" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.110832 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx"] Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.244300 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl2xb\" (UniqueName: \"kubernetes.io/projected/311ed4a0-c6f5-459d-9da6-86dbe84dd12f-kube-api-access-jl2xb\") pod \"nmstate-operator-5d6f6cfd66-hh9gx\" (UID: \"311ed4a0-c6f5-459d-9da6-86dbe84dd12f\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.345617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl2xb\" (UniqueName: \"kubernetes.io/projected/311ed4a0-c6f5-459d-9da6-86dbe84dd12f-kube-api-access-jl2xb\") pod \"nmstate-operator-5d6f6cfd66-hh9gx\" (UID: \"311ed4a0-c6f5-459d-9da6-86dbe84dd12f\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.365458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl2xb\" (UniqueName: \"kubernetes.io/projected/311ed4a0-c6f5-459d-9da6-86dbe84dd12f-kube-api-access-jl2xb\") pod \"nmstate-operator-5d6f6cfd66-hh9gx\" (UID: \"311ed4a0-c6f5-459d-9da6-86dbe84dd12f\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.407123 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx" Sep 30 14:32:35 crc kubenswrapper[4799]: I0930 14:32:35.952227 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx"] Sep 30 14:32:35 crc kubenswrapper[4799]: W0930 14:32:35.970434 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod311ed4a0_c6f5_459d_9da6_86dbe84dd12f.slice/crio-b2d852b1da44a7b8edbadcf45c08dbb68e7c84bd97d537fb87b55d56bc5af70f WatchSource:0}: Error finding container b2d852b1da44a7b8edbadcf45c08dbb68e7c84bd97d537fb87b55d56bc5af70f: Status 404 returned error can't find the container with id b2d852b1da44a7b8edbadcf45c08dbb68e7c84bd97d537fb87b55d56bc5af70f Sep 30 14:32:36 crc kubenswrapper[4799]: I0930 14:32:36.144022 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:36 crc kubenswrapper[4799]: I0930 14:32:36.144084 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:36 crc kubenswrapper[4799]: I0930 14:32:36.210467 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:36 crc kubenswrapper[4799]: I0930 14:32:36.608474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx" event={"ID":"311ed4a0-c6f5-459d-9da6-86dbe84dd12f","Type":"ContainerStarted","Data":"b2d852b1da44a7b8edbadcf45c08dbb68e7c84bd97d537fb87b55d56bc5af70f"} Sep 30 14:32:36 crc kubenswrapper[4799]: I0930 14:32:36.655105 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:38 crc kubenswrapper[4799]: I0930 14:32:38.612770 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mp8vs"] Sep 30 14:32:38 crc kubenswrapper[4799]: I0930 14:32:38.630520 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mp8vs" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerName="registry-server" containerID="cri-o://7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2" gracePeriod=2 Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.155926 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.302218 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz6zp\" (UniqueName: \"kubernetes.io/projected/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-kube-api-access-dz6zp\") pod \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.302404 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-catalog-content\") pod \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.302440 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-utilities\") pod \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\" (UID: \"2ed54eba-fa9c-4754-82c1-d697d9bc40b8\") " Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.303548 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-utilities" (OuterVolumeSpecName: "utilities") pod "2ed54eba-fa9c-4754-82c1-d697d9bc40b8" (UID: "2ed54eba-fa9c-4754-82c1-d697d9bc40b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.307295 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-kube-api-access-dz6zp" (OuterVolumeSpecName: "kube-api-access-dz6zp") pod "2ed54eba-fa9c-4754-82c1-d697d9bc40b8" (UID: "2ed54eba-fa9c-4754-82c1-d697d9bc40b8"). InnerVolumeSpecName "kube-api-access-dz6zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.391954 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ed54eba-fa9c-4754-82c1-d697d9bc40b8" (UID: "2ed54eba-fa9c-4754-82c1-d697d9bc40b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.403889 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz6zp\" (UniqueName: \"kubernetes.io/projected/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-kube-api-access-dz6zp\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.403936 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.403946 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed54eba-fa9c-4754-82c1-d697d9bc40b8-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.641933 4799 generic.go:334] "Generic (PLEG): container finished" podID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerID="7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2" exitCode=0 Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.642050 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mp8vs" event={"ID":"2ed54eba-fa9c-4754-82c1-d697d9bc40b8","Type":"ContainerDied","Data":"7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2"} Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.642068 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mp8vs" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.642546 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mp8vs" event={"ID":"2ed54eba-fa9c-4754-82c1-d697d9bc40b8","Type":"ContainerDied","Data":"64fec58d9ee16c59f66fb5a1a3fba1da89a0e0352a57339c6aa676991f1577e4"} Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.642588 4799 scope.go:117] "RemoveContainer" containerID="7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.647613 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx" event={"ID":"311ed4a0-c6f5-459d-9da6-86dbe84dd12f","Type":"ContainerStarted","Data":"590c6ced16501b3cf25e7ba5df9ac9a258bae3b50563272760925d8a71bf929b"} Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.677272 4799 scope.go:117] "RemoveContainer" containerID="8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.690594 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-hh9gx" podStartSLOduration=2.188142514 podStartE2EDuration="4.690573659s" podCreationTimestamp="2025-09-30 14:32:35 +0000 UTC" firstStartedPulling="2025-09-30 14:32:35.975998655 +0000 UTC m=+778.059599082" lastFinishedPulling="2025-09-30 14:32:38.47842981 +0000 UTC m=+780.562030227" observedRunningTime="2025-09-30 14:32:39.671157832 +0000 UTC m=+781.754758279" watchObservedRunningTime="2025-09-30 14:32:39.690573659 +0000 UTC m=+781.774174086" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.699766 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mp8vs"] Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.702685 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mp8vs"] Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.723028 4799 scope.go:117] "RemoveContainer" containerID="da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.741402 4799 scope.go:117] "RemoveContainer" containerID="7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2" Sep 30 14:32:39 crc kubenswrapper[4799]: E0930 14:32:39.742098 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2\": container with ID starting with 7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2 not found: ID does not exist" containerID="7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.742131 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2"} err="failed to get container status \"7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2\": rpc error: code = NotFound desc = could not find container \"7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2\": container with ID starting with 7acce2778bea0934eb348b5fe9fc5a07008929136285ef2e466292b77ffed6c2 not found: ID does not exist" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.742157 4799 scope.go:117] "RemoveContainer" containerID="8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b" Sep 30 14:32:39 crc kubenswrapper[4799]: E0930 14:32:39.742880 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b\": container with ID starting with 8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b not found: ID does not exist" containerID="8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.742930 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b"} err="failed to get container status \"8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b\": rpc error: code = NotFound desc = could not find container \"8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b\": container with ID starting with 8a9c53c5a5dc7adf71302f72b0681b6c1005187efc87adae6c804d092a52303b not found: ID does not exist" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.742948 4799 scope.go:117] "RemoveContainer" containerID="da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f" Sep 30 14:32:39 crc kubenswrapper[4799]: E0930 14:32:39.743237 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f\": container with ID starting with da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f not found: ID does not exist" containerID="da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f" Sep 30 14:32:39 crc kubenswrapper[4799]: I0930 14:32:39.743260 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f"} err="failed to get container status \"da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f\": rpc error: code = NotFound desc = could not find container \"da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f\": container with ID starting with da6d576fc6883f271ae0174d20a83772190767cc71ec2357831e8aab8599541f not found: ID does not exist" Sep 30 14:32:40 crc kubenswrapper[4799]: I0930 14:32:40.510385 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" path="/var/lib/kubelet/pods/2ed54eba-fa9c-4754-82c1-d697d9bc40b8/volumes" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.112291 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx"] Sep 30 14:32:44 crc kubenswrapper[4799]: E0930 14:32:44.112957 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerName="extract-utilities" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.112973 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerName="extract-utilities" Sep 30 14:32:44 crc kubenswrapper[4799]: E0930 14:32:44.112987 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerName="registry-server" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.112997 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerName="registry-server" Sep 30 14:32:44 crc kubenswrapper[4799]: E0930 14:32:44.113018 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerName="extract-content" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.113025 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerName="extract-content" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.113142 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ed54eba-fa9c-4754-82c1-d697d9bc40b8" containerName="registry-server" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.113853 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.116548 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-c5grk" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.128387 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx"] Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.143325 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-jdngw"] Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.144222 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.147567 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.170847 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-jdngw"] Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.174560 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-672fh"] Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.175338 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.264352 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-nmstate-lock\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.264410 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlx45\" (UniqueName: \"kubernetes.io/projected/dc69b189-0b92-42fb-9584-42fd5aed183f-kube-api-access-zlx45\") pod \"nmstate-metrics-58fcddf996-bk8lx\" (UID: \"dc69b189-0b92-42fb-9584-42fd5aed183f\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.264434 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-dbus-socket\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.264473 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqw2p\" (UniqueName: \"kubernetes.io/projected/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-kube-api-access-hqw2p\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.264752 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-ovs-socket\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.264857 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5bwd\" (UniqueName: \"kubernetes.io/projected/23e26176-f572-43f8-baba-8bbe27fa0bca-kube-api-access-x5bwd\") pod \"nmstate-webhook-6d689559c5-jdngw\" (UID: \"23e26176-f572-43f8-baba-8bbe27fa0bca\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.264929 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/23e26176-f572-43f8-baba-8bbe27fa0bca-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-jdngw\" (UID: \"23e26176-f572-43f8-baba-8bbe27fa0bca\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.298326 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f"] Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.299003 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.300974 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.301936 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.307340 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-wwjbh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.321684 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f"] Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.365970 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-nmstate-lock\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.366025 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlx45\" (UniqueName: \"kubernetes.io/projected/dc69b189-0b92-42fb-9584-42fd5aed183f-kube-api-access-zlx45\") pod \"nmstate-metrics-58fcddf996-bk8lx\" (UID: \"dc69b189-0b92-42fb-9584-42fd5aed183f\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.366051 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-dbus-socket\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.366076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqw2p\" (UniqueName: \"kubernetes.io/projected/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-kube-api-access-hqw2p\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.366110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-ovs-socket\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.366161 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5bwd\" (UniqueName: \"kubernetes.io/projected/23e26176-f572-43f8-baba-8bbe27fa0bca-kube-api-access-x5bwd\") pod \"nmstate-webhook-6d689559c5-jdngw\" (UID: \"23e26176-f572-43f8-baba-8bbe27fa0bca\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.366197 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/23e26176-f572-43f8-baba-8bbe27fa0bca-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-jdngw\" (UID: \"23e26176-f572-43f8-baba-8bbe27fa0bca\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.366406 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-dbus-socket\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.366530 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-ovs-socket\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.367004 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-nmstate-lock\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.393815 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqw2p\" (UniqueName: \"kubernetes.io/projected/60de70ea-f8ce-4f5a-ab69-45fe7cd77e20-kube-api-access-hqw2p\") pod \"nmstate-handler-672fh\" (UID: \"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20\") " pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.396598 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/23e26176-f572-43f8-baba-8bbe27fa0bca-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-jdngw\" (UID: \"23e26176-f572-43f8-baba-8bbe27fa0bca\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.399416 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5bwd\" (UniqueName: \"kubernetes.io/projected/23e26176-f572-43f8-baba-8bbe27fa0bca-kube-api-access-x5bwd\") pod \"nmstate-webhook-6d689559c5-jdngw\" (UID: \"23e26176-f572-43f8-baba-8bbe27fa0bca\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.399451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlx45\" (UniqueName: \"kubernetes.io/projected/dc69b189-0b92-42fb-9584-42fd5aed183f-kube-api-access-zlx45\") pod \"nmstate-metrics-58fcddf996-bk8lx\" (UID: \"dc69b189-0b92-42fb-9584-42fd5aed183f\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.431681 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.467149 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.468894 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.472984 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm9j9\" (UniqueName: \"kubernetes.io/projected/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-kube-api-access-nm9j9\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.473214 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.488414 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.555685 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5d57b4bfbb-7rw7q"] Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.556908 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.574721 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.575841 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.575968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm9j9\" (UniqueName: \"kubernetes.io/projected/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-kube-api-access-nm9j9\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.575513 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5d57b4bfbb-7rw7q"] Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.576520 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: E0930 14:32:44.576573 4799 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Sep 30 14:32:44 crc kubenswrapper[4799]: E0930 14:32:44.576809 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-plugin-serving-cert podName:3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69 nodeName:}" failed. No retries permitted until 2025-09-30 14:32:45.076794583 +0000 UTC m=+787.160395010 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-plugin-serving-cert") pod "nmstate-console-plugin-864bb6dfb5-pbz2f" (UID: "3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69") : secret "plugin-serving-cert" not found Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.649790 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm9j9\" (UniqueName: \"kubernetes.io/projected/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-kube-api-access-nm9j9\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.677363 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv5dh\" (UniqueName: \"kubernetes.io/projected/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-kube-api-access-gv5dh\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.677399 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-trusted-ca-bundle\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.677428 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-oauth-serving-cert\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.677446 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-config\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.677590 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-serving-cert\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.677790 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-oauth-config\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.677867 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-service-ca\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.706975 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-672fh" event={"ID":"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20","Type":"ContainerStarted","Data":"89e77c0a1a23dbe98963c103e083fa199ddac0df2fb9ae14d0ea9aa34863ff7f"} Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.779346 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-serving-cert\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.779456 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-oauth-config\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.779478 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-service-ca\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.779511 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv5dh\" (UniqueName: \"kubernetes.io/projected/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-kube-api-access-gv5dh\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.779532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-trusted-ca-bundle\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.779554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-oauth-serving-cert\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.779578 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-config\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.780595 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-config\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.782022 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-trusted-ca-bundle\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.783519 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-service-ca\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.783681 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-oauth-serving-cert\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.786263 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-serving-cert\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.786840 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-console-oauth-config\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.802518 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv5dh\" (UniqueName: \"kubernetes.io/projected/db22eaf6-5b5d-4a20-9cec-d64847faa2fd-kube-api-access-gv5dh\") pod \"console-5d57b4bfbb-7rw7q\" (UID: \"db22eaf6-5b5d-4a20-9cec-d64847faa2fd\") " pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.848920 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-jdngw"] Sep 30 14:32:44 crc kubenswrapper[4799]: W0930 14:32:44.857082 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23e26176_f572_43f8_baba_8bbe27fa0bca.slice/crio-f0593b1094c052d56b1198b104c506cdc0428666e1a0f8f0806aa33a5d54703d WatchSource:0}: Error finding container f0593b1094c052d56b1198b104c506cdc0428666e1a0f8f0806aa33a5d54703d: Status 404 returned error can't find the container with id f0593b1094c052d56b1198b104c506cdc0428666e1a0f8f0806aa33a5d54703d Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.872250 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:44 crc kubenswrapper[4799]: I0930 14:32:44.977122 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx"] Sep 30 14:32:44 crc kubenswrapper[4799]: W0930 14:32:44.987902 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc69b189_0b92_42fb_9584_42fd5aed183f.slice/crio-2c207697ac035e216d62d4c13ffa560ac2a625e2a775be7ad7dfaf9b748e1682 WatchSource:0}: Error finding container 2c207697ac035e216d62d4c13ffa560ac2a625e2a775be7ad7dfaf9b748e1682: Status 404 returned error can't find the container with id 2c207697ac035e216d62d4c13ffa560ac2a625e2a775be7ad7dfaf9b748e1682 Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.088121 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.093312 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-pbz2f\" (UID: \"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.214185 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.327034 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5d57b4bfbb-7rw7q"] Sep 30 14:32:45 crc kubenswrapper[4799]: W0930 14:32:45.332924 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb22eaf6_5b5d_4a20_9cec_d64847faa2fd.slice/crio-9c99e38b18f206c5fac550b9a22d764f1c0904e58429770fb5fec8c07883b068 WatchSource:0}: Error finding container 9c99e38b18f206c5fac550b9a22d764f1c0904e58429770fb5fec8c07883b068: Status 404 returned error can't find the container with id 9c99e38b18f206c5fac550b9a22d764f1c0904e58429770fb5fec8c07883b068 Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.616290 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f"] Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.713475 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" event={"ID":"dc69b189-0b92-42fb-9584-42fd5aed183f","Type":"ContainerStarted","Data":"2c207697ac035e216d62d4c13ffa560ac2a625e2a775be7ad7dfaf9b748e1682"} Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.714763 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" event={"ID":"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69","Type":"ContainerStarted","Data":"5192173d74532aa62add69c39f2c8bf5ea72d8c6bdd20d6f36f31e8d103c0550"} Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.716380 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" event={"ID":"23e26176-f572-43f8-baba-8bbe27fa0bca","Type":"ContainerStarted","Data":"f0593b1094c052d56b1198b104c506cdc0428666e1a0f8f0806aa33a5d54703d"} Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.718719 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d57b4bfbb-7rw7q" event={"ID":"db22eaf6-5b5d-4a20-9cec-d64847faa2fd","Type":"ContainerStarted","Data":"dd89350d885c1342f1b0d0e9aba1ae84f3bd7843d6808a575544455c83aa7ab9"} Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.718778 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d57b4bfbb-7rw7q" event={"ID":"db22eaf6-5b5d-4a20-9cec-d64847faa2fd","Type":"ContainerStarted","Data":"9c99e38b18f206c5fac550b9a22d764f1c0904e58429770fb5fec8c07883b068"} Sep 30 14:32:45 crc kubenswrapper[4799]: I0930 14:32:45.744187 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5d57b4bfbb-7rw7q" podStartSLOduration=1.744163755 podStartE2EDuration="1.744163755s" podCreationTimestamp="2025-09-30 14:32:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:32:45.739104917 +0000 UTC m=+787.822705354" watchObservedRunningTime="2025-09-30 14:32:45.744163755 +0000 UTC m=+787.827764182" Sep 30 14:32:48 crc kubenswrapper[4799]: I0930 14:32:48.743980 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" event={"ID":"dc69b189-0b92-42fb-9584-42fd5aed183f","Type":"ContainerStarted","Data":"fbebef9516a46c6eaced9beab853af2ba60ee448708a94ada43931bce85172d4"} Sep 30 14:32:48 crc kubenswrapper[4799]: I0930 14:32:48.745941 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" event={"ID":"23e26176-f572-43f8-baba-8bbe27fa0bca","Type":"ContainerStarted","Data":"43a1c4b7b2a660a07eb042c555980bccff02643665b4d94b742f5afd9c103b6d"} Sep 30 14:32:48 crc kubenswrapper[4799]: I0930 14:32:48.746124 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:32:48 crc kubenswrapper[4799]: I0930 14:32:48.747985 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-672fh" event={"ID":"60de70ea-f8ce-4f5a-ab69-45fe7cd77e20","Type":"ContainerStarted","Data":"a837572f043b143dc6ea71c28d3b879d45779ac827b0b141394c5d91d7ce57b1"} Sep 30 14:32:48 crc kubenswrapper[4799]: I0930 14:32:48.748329 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:48 crc kubenswrapper[4799]: I0930 14:32:48.768248 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" podStartSLOduration=1.8840509 podStartE2EDuration="4.768221204s" podCreationTimestamp="2025-09-30 14:32:44 +0000 UTC" firstStartedPulling="2025-09-30 14:32:44.861251358 +0000 UTC m=+786.944851785" lastFinishedPulling="2025-09-30 14:32:47.745421662 +0000 UTC m=+789.829022089" observedRunningTime="2025-09-30 14:32:48.768050149 +0000 UTC m=+790.851650596" watchObservedRunningTime="2025-09-30 14:32:48.768221204 +0000 UTC m=+790.851821631" Sep 30 14:32:48 crc kubenswrapper[4799]: I0930 14:32:48.799721 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-672fh" podStartSLOduration=1.613931183 podStartE2EDuration="4.799693712s" podCreationTimestamp="2025-09-30 14:32:44 +0000 UTC" firstStartedPulling="2025-09-30 14:32:44.536188518 +0000 UTC m=+786.619788945" lastFinishedPulling="2025-09-30 14:32:47.721951047 +0000 UTC m=+789.805551474" observedRunningTime="2025-09-30 14:32:48.795870621 +0000 UTC m=+790.879471078" watchObservedRunningTime="2025-09-30 14:32:48.799693712 +0000 UTC m=+790.883294159" Sep 30 14:32:49 crc kubenswrapper[4799]: I0930 14:32:49.759222 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" event={"ID":"3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69","Type":"ContainerStarted","Data":"594bc3fedac6b952de12ded2d8eba25001620e6316f404b2e126c6a90c893d6f"} Sep 30 14:32:49 crc kubenswrapper[4799]: I0930 14:32:49.783198 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-pbz2f" podStartSLOduration=2.1417704029999998 podStartE2EDuration="5.783174855s" podCreationTimestamp="2025-09-30 14:32:44 +0000 UTC" firstStartedPulling="2025-09-30 14:32:45.626280983 +0000 UTC m=+787.709881400" lastFinishedPulling="2025-09-30 14:32:49.267685435 +0000 UTC m=+791.351285852" observedRunningTime="2025-09-30 14:32:49.781041052 +0000 UTC m=+791.864641499" watchObservedRunningTime="2025-09-30 14:32:49.783174855 +0000 UTC m=+791.866775292" Sep 30 14:32:51 crc kubenswrapper[4799]: I0930 14:32:51.771637 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" event={"ID":"dc69b189-0b92-42fb-9584-42fd5aed183f","Type":"ContainerStarted","Data":"ef1faadbf88e45c042572c5b01bf0263282d769e46e4e4614c4a1ea364e385d2"} Sep 30 14:32:51 crc kubenswrapper[4799]: I0930 14:32:51.793823 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-bk8lx" podStartSLOduration=2.019048121 podStartE2EDuration="7.793803496s" podCreationTimestamp="2025-09-30 14:32:44 +0000 UTC" firstStartedPulling="2025-09-30 14:32:44.990814811 +0000 UTC m=+787.074415238" lastFinishedPulling="2025-09-30 14:32:50.765570186 +0000 UTC m=+792.849170613" observedRunningTime="2025-09-30 14:32:51.789183971 +0000 UTC m=+793.872784418" watchObservedRunningTime="2025-09-30 14:32:51.793803496 +0000 UTC m=+793.877403923" Sep 30 14:32:54 crc kubenswrapper[4799]: I0930 14:32:54.511396 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-672fh" Sep 30 14:32:54 crc kubenswrapper[4799]: I0930 14:32:54.873309 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:54 crc kubenswrapper[4799]: I0930 14:32:54.873929 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:54 crc kubenswrapper[4799]: I0930 14:32:54.878097 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:55 crc kubenswrapper[4799]: I0930 14:32:55.801859 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5d57b4bfbb-7rw7q" Sep 30 14:32:55 crc kubenswrapper[4799]: I0930 14:32:55.862934 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-svk57"] Sep 30 14:33:00 crc kubenswrapper[4799]: I0930 14:33:00.929073 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dt97k"] Sep 30 14:33:00 crc kubenswrapper[4799]: I0930 14:33:00.932368 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:00 crc kubenswrapper[4799]: I0930 14:33:00.944826 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dt97k"] Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.018299 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j859m\" (UniqueName: \"kubernetes.io/projected/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-kube-api-access-j859m\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.018414 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-utilities\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.018490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-catalog-content\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.119591 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-catalog-content\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.119702 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j859m\" (UniqueName: \"kubernetes.io/projected/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-kube-api-access-j859m\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.119730 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-utilities\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.120316 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-catalog-content\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.120335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-utilities\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.142800 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j859m\" (UniqueName: \"kubernetes.io/projected/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-kube-api-access-j859m\") pod \"redhat-marketplace-dt97k\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.259285 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.544572 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dt97k"] Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.831128 4799 generic.go:334] "Generic (PLEG): container finished" podID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerID="4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024" exitCode=0 Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.831329 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dt97k" event={"ID":"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb","Type":"ContainerDied","Data":"4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024"} Sep 30 14:33:01 crc kubenswrapper[4799]: I0930 14:33:01.832086 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dt97k" event={"ID":"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb","Type":"ContainerStarted","Data":"629d91d24697a80b279b3e5723cedf7e13d9e46383bbd4366855611dfc7d3892"} Sep 30 14:33:03 crc kubenswrapper[4799]: I0930 14:33:03.843262 4799 generic.go:334] "Generic (PLEG): container finished" podID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerID="5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5" exitCode=0 Sep 30 14:33:03 crc kubenswrapper[4799]: I0930 14:33:03.843335 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dt97k" event={"ID":"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb","Type":"ContainerDied","Data":"5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5"} Sep 30 14:33:04 crc kubenswrapper[4799]: I0930 14:33:04.475090 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-jdngw" Sep 30 14:33:04 crc kubenswrapper[4799]: I0930 14:33:04.851465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dt97k" event={"ID":"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb","Type":"ContainerStarted","Data":"6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29"} Sep 30 14:33:04 crc kubenswrapper[4799]: I0930 14:33:04.873717 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dt97k" podStartSLOduration=2.406100203 podStartE2EDuration="4.873696665s" podCreationTimestamp="2025-09-30 14:33:00 +0000 UTC" firstStartedPulling="2025-09-30 14:33:01.834435333 +0000 UTC m=+803.918035760" lastFinishedPulling="2025-09-30 14:33:04.302031795 +0000 UTC m=+806.385632222" observedRunningTime="2025-09-30 14:33:04.870844782 +0000 UTC m=+806.954445219" watchObservedRunningTime="2025-09-30 14:33:04.873696665 +0000 UTC m=+806.957297102" Sep 30 14:33:11 crc kubenswrapper[4799]: I0930 14:33:11.260101 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:11 crc kubenswrapper[4799]: I0930 14:33:11.260946 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:11 crc kubenswrapper[4799]: I0930 14:33:11.301385 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:11 crc kubenswrapper[4799]: I0930 14:33:11.958861 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:12 crc kubenswrapper[4799]: I0930 14:33:12.058154 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dt97k"] Sep 30 14:33:13 crc kubenswrapper[4799]: I0930 14:33:13.908788 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dt97k" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerName="registry-server" containerID="cri-o://6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29" gracePeriod=2 Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.302492 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.426682 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j859m\" (UniqueName: \"kubernetes.io/projected/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-kube-api-access-j859m\") pod \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.426724 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-catalog-content\") pod \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.426775 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-utilities\") pod \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\" (UID: \"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb\") " Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.428182 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-utilities" (OuterVolumeSpecName: "utilities") pod "ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" (UID: "ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.434282 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-kube-api-access-j859m" (OuterVolumeSpecName: "kube-api-access-j859m") pod "ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" (UID: "ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb"). InnerVolumeSpecName "kube-api-access-j859m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.447351 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" (UID: "ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.528542 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j859m\" (UniqueName: \"kubernetes.io/projected/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-kube-api-access-j859m\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.528582 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.528592 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.916753 4799 generic.go:334] "Generic (PLEG): container finished" podID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerID="6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29" exitCode=0 Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.916803 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dt97k" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.916830 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dt97k" event={"ID":"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb","Type":"ContainerDied","Data":"6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29"} Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.917155 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dt97k" event={"ID":"ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb","Type":"ContainerDied","Data":"629d91d24697a80b279b3e5723cedf7e13d9e46383bbd4366855611dfc7d3892"} Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.917186 4799 scope.go:117] "RemoveContainer" containerID="6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.948077 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dt97k"] Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.954004 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dt97k"] Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.961256 4799 scope.go:117] "RemoveContainer" containerID="5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5" Sep 30 14:33:14 crc kubenswrapper[4799]: I0930 14:33:14.978220 4799 scope.go:117] "RemoveContainer" containerID="4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024" Sep 30 14:33:15 crc kubenswrapper[4799]: I0930 14:33:15.007248 4799 scope.go:117] "RemoveContainer" containerID="6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29" Sep 30 14:33:15 crc kubenswrapper[4799]: E0930 14:33:15.007807 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29\": container with ID starting with 6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29 not found: ID does not exist" containerID="6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29" Sep 30 14:33:15 crc kubenswrapper[4799]: I0930 14:33:15.007863 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29"} err="failed to get container status \"6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29\": rpc error: code = NotFound desc = could not find container \"6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29\": container with ID starting with 6af93756ba369706a30ad066fa35d52c5e527ea9d10c16126e73674ef6af0a29 not found: ID does not exist" Sep 30 14:33:15 crc kubenswrapper[4799]: I0930 14:33:15.007900 4799 scope.go:117] "RemoveContainer" containerID="5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5" Sep 30 14:33:15 crc kubenswrapper[4799]: E0930 14:33:15.008246 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5\": container with ID starting with 5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5 not found: ID does not exist" containerID="5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5" Sep 30 14:33:15 crc kubenswrapper[4799]: I0930 14:33:15.008315 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5"} err="failed to get container status \"5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5\": rpc error: code = NotFound desc = could not find container \"5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5\": container with ID starting with 5bee50eedfad30df0069ba4d4a12614f664828a4a128be744c8e0de3e94514c5 not found: ID does not exist" Sep 30 14:33:15 crc kubenswrapper[4799]: I0930 14:33:15.008340 4799 scope.go:117] "RemoveContainer" containerID="4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024" Sep 30 14:33:15 crc kubenswrapper[4799]: E0930 14:33:15.008770 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024\": container with ID starting with 4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024 not found: ID does not exist" containerID="4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024" Sep 30 14:33:15 crc kubenswrapper[4799]: I0930 14:33:15.008816 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024"} err="failed to get container status \"4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024\": rpc error: code = NotFound desc = could not find container \"4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024\": container with ID starting with 4f5319c7061e5ac87ef2f7066ef5c37652fcebd805247d6720943315a0e99024 not found: ID does not exist" Sep 30 14:33:16 crc kubenswrapper[4799]: I0930 14:33:16.515037 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" path="/var/lib/kubelet/pods/ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb/volumes" Sep 30 14:33:16 crc kubenswrapper[4799]: I0930 14:33:16.939229 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cr57c"] Sep 30 14:33:16 crc kubenswrapper[4799]: E0930 14:33:16.939523 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerName="registry-server" Sep 30 14:33:16 crc kubenswrapper[4799]: I0930 14:33:16.939535 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerName="registry-server" Sep 30 14:33:16 crc kubenswrapper[4799]: E0930 14:33:16.939560 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerName="extract-utilities" Sep 30 14:33:16 crc kubenswrapper[4799]: I0930 14:33:16.939566 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerName="extract-utilities" Sep 30 14:33:16 crc kubenswrapper[4799]: E0930 14:33:16.939576 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerName="extract-content" Sep 30 14:33:16 crc kubenswrapper[4799]: I0930 14:33:16.939582 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerName="extract-content" Sep 30 14:33:16 crc kubenswrapper[4799]: I0930 14:33:16.940791 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae25f2c8-ab00-4207-aaa5-0de3d62b0ccb" containerName="registry-server" Sep 30 14:33:16 crc kubenswrapper[4799]: I0930 14:33:16.941869 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:16 crc kubenswrapper[4799]: I0930 14:33:16.958864 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cr57c"] Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.077744 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4jwd\" (UniqueName: \"kubernetes.io/projected/185405f2-a875-49de-9107-ec4772cc69a5-kube-api-access-n4jwd\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.077811 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-utilities\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.077838 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-catalog-content\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.178791 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4jwd\" (UniqueName: \"kubernetes.io/projected/185405f2-a875-49de-9107-ec4772cc69a5-kube-api-access-n4jwd\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.179121 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-utilities\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.179145 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-catalog-content\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.179782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-utilities\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.179805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-catalog-content\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.201044 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4jwd\" (UniqueName: \"kubernetes.io/projected/185405f2-a875-49de-9107-ec4772cc69a5-kube-api-access-n4jwd\") pod \"community-operators-cr57c\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:17 crc kubenswrapper[4799]: I0930 14:33:17.265889 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:17.605807 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cr57c"] Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:17.943503 4799 generic.go:334] "Generic (PLEG): container finished" podID="185405f2-a875-49de-9107-ec4772cc69a5" containerID="f7486679dc40c58b850350509c501d6e3384f62f5908997248ebd44cab0a5779" exitCode=0 Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:17.943542 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr57c" event={"ID":"185405f2-a875-49de-9107-ec4772cc69a5","Type":"ContainerDied","Data":"f7486679dc40c58b850350509c501d6e3384f62f5908997248ebd44cab0a5779"} Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:17.943570 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr57c" event={"ID":"185405f2-a875-49de-9107-ec4772cc69a5","Type":"ContainerStarted","Data":"28470333e857e4ab6c3c983b1812f68a65ee6c14e1bc4c4b0ad6d3942d7acf7f"} Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:18.801890 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b"] Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:18.803527 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:18.805464 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:18.823414 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b"] Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:18.903852 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:18.904224 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:18.904308 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9njk\" (UniqueName: \"kubernetes.io/projected/90186c62-dccd-4a14-908c-6b6b5c1750ba-kube-api-access-m9njk\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:18 crc kubenswrapper[4799]: I0930 14:33:18.957066 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr57c" event={"ID":"185405f2-a875-49de-9107-ec4772cc69a5","Type":"ContainerStarted","Data":"0401d2f91c708e82578e9b1bb8d3fe561e6431d843d0d779de6dce29bd22f269"} Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.006290 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.006356 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.006426 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9njk\" (UniqueName: \"kubernetes.io/projected/90186c62-dccd-4a14-908c-6b6b5c1750ba-kube-api-access-m9njk\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.007421 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.007459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.032056 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9njk\" (UniqueName: \"kubernetes.io/projected/90186c62-dccd-4a14-908c-6b6b5c1750ba-kube-api-access-m9njk\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.123864 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.627089 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b"] Sep 30 14:33:19 crc kubenswrapper[4799]: W0930 14:33:19.634015 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90186c62_dccd_4a14_908c_6b6b5c1750ba.slice/crio-362e2ea46d11834aa1e6fa1667f386c9da1439eec21d89687360381199f93248 WatchSource:0}: Error finding container 362e2ea46d11834aa1e6fa1667f386c9da1439eec21d89687360381199f93248: Status 404 returned error can't find the container with id 362e2ea46d11834aa1e6fa1667f386c9da1439eec21d89687360381199f93248 Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.962661 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" event={"ID":"90186c62-dccd-4a14-908c-6b6b5c1750ba","Type":"ContainerStarted","Data":"362e2ea46d11834aa1e6fa1667f386c9da1439eec21d89687360381199f93248"} Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.964296 4799 generic.go:334] "Generic (PLEG): container finished" podID="185405f2-a875-49de-9107-ec4772cc69a5" containerID="0401d2f91c708e82578e9b1bb8d3fe561e6431d843d0d779de6dce29bd22f269" exitCode=0 Sep 30 14:33:19 crc kubenswrapper[4799]: I0930 14:33:19.964330 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr57c" event={"ID":"185405f2-a875-49de-9107-ec4772cc69a5","Type":"ContainerDied","Data":"0401d2f91c708e82578e9b1bb8d3fe561e6431d843d0d779de6dce29bd22f269"} Sep 30 14:33:20 crc kubenswrapper[4799]: I0930 14:33:20.907821 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-svk57" podUID="03c9ee3f-722e-4db8-a8f5-dc359486e4f4" containerName="console" containerID="cri-o://0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115" gracePeriod=15 Sep 30 14:33:20 crc kubenswrapper[4799]: I0930 14:33:20.973181 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr57c" event={"ID":"185405f2-a875-49de-9107-ec4772cc69a5","Type":"ContainerStarted","Data":"a1f564328632914ce92ac09d826d628835491de2713fec5554c6e3a86e3e5401"} Sep 30 14:33:20 crc kubenswrapper[4799]: I0930 14:33:20.975753 4799 generic.go:334] "Generic (PLEG): container finished" podID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerID="ba307df50474a39622bdc6dc55c26b6d5e7926982b29c64783ee60858ccc704b" exitCode=0 Sep 30 14:33:20 crc kubenswrapper[4799]: I0930 14:33:20.975795 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" event={"ID":"90186c62-dccd-4a14-908c-6b6b5c1750ba","Type":"ContainerDied","Data":"ba307df50474a39622bdc6dc55c26b6d5e7926982b29c64783ee60858ccc704b"} Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.002542 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cr57c" podStartSLOduration=2.369067916 podStartE2EDuration="5.002517141s" podCreationTimestamp="2025-09-30 14:33:16 +0000 UTC" firstStartedPulling="2025-09-30 14:33:17.944796749 +0000 UTC m=+820.028397176" lastFinishedPulling="2025-09-30 14:33:20.578245974 +0000 UTC m=+822.661846401" observedRunningTime="2025-09-30 14:33:20.997705731 +0000 UTC m=+823.081306188" watchObservedRunningTime="2025-09-30 14:33:21.002517141 +0000 UTC m=+823.086117568" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.312707 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-svk57_03c9ee3f-722e-4db8-a8f5-dc359486e4f4/console/0.log" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.312963 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.438826 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-oauth-config\") pod \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.438899 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm2jz\" (UniqueName: \"kubernetes.io/projected/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-kube-api-access-rm2jz\") pod \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.438932 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-config\") pod \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.438956 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-trusted-ca-bundle\") pod \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.438971 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-service-ca\") pod \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.439020 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-serving-cert\") pod \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.439070 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-oauth-serving-cert\") pod \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\" (UID: \"03c9ee3f-722e-4db8-a8f5-dc359486e4f4\") " Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.440466 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "03c9ee3f-722e-4db8-a8f5-dc359486e4f4" (UID: "03c9ee3f-722e-4db8-a8f5-dc359486e4f4"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.440559 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-config" (OuterVolumeSpecName: "console-config") pod "03c9ee3f-722e-4db8-a8f5-dc359486e4f4" (UID: "03c9ee3f-722e-4db8-a8f5-dc359486e4f4"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.441181 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-service-ca" (OuterVolumeSpecName: "service-ca") pod "03c9ee3f-722e-4db8-a8f5-dc359486e4f4" (UID: "03c9ee3f-722e-4db8-a8f5-dc359486e4f4"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.441243 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "03c9ee3f-722e-4db8-a8f5-dc359486e4f4" (UID: "03c9ee3f-722e-4db8-a8f5-dc359486e4f4"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.445606 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "03c9ee3f-722e-4db8-a8f5-dc359486e4f4" (UID: "03c9ee3f-722e-4db8-a8f5-dc359486e4f4"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.445714 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-kube-api-access-rm2jz" (OuterVolumeSpecName: "kube-api-access-rm2jz") pod "03c9ee3f-722e-4db8-a8f5-dc359486e4f4" (UID: "03c9ee3f-722e-4db8-a8f5-dc359486e4f4"). InnerVolumeSpecName "kube-api-access-rm2jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.446607 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "03c9ee3f-722e-4db8-a8f5-dc359486e4f4" (UID: "03c9ee3f-722e-4db8-a8f5-dc359486e4f4"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.541302 4799 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.541362 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm2jz\" (UniqueName: \"kubernetes.io/projected/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-kube-api-access-rm2jz\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.541376 4799 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.541387 4799 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.541402 4799 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-service-ca\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.541412 4799 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.541431 4799 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03c9ee3f-722e-4db8-a8f5-dc359486e4f4-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.992753 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-svk57_03c9ee3f-722e-4db8-a8f5-dc359486e4f4/console/0.log" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.992860 4799 generic.go:334] "Generic (PLEG): container finished" podID="03c9ee3f-722e-4db8-a8f5-dc359486e4f4" containerID="0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115" exitCode=2 Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.993335 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-svk57" event={"ID":"03c9ee3f-722e-4db8-a8f5-dc359486e4f4","Type":"ContainerDied","Data":"0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115"} Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.993413 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-svk57" event={"ID":"03c9ee3f-722e-4db8-a8f5-dc359486e4f4","Type":"ContainerDied","Data":"ebeb1f8e458db03a90e455eefe8c04d58595ba5b8d4243c8b201320a2187aba8"} Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.993438 4799 scope.go:117] "RemoveContainer" containerID="0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115" Sep 30 14:33:21 crc kubenswrapper[4799]: I0930 14:33:21.993706 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-svk57" Sep 30 14:33:22 crc kubenswrapper[4799]: I0930 14:33:22.019983 4799 scope.go:117] "RemoveContainer" containerID="0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115" Sep 30 14:33:22 crc kubenswrapper[4799]: E0930 14:33:22.021195 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115\": container with ID starting with 0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115 not found: ID does not exist" containerID="0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115" Sep 30 14:33:22 crc kubenswrapper[4799]: I0930 14:33:22.021234 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115"} err="failed to get container status \"0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115\": rpc error: code = NotFound desc = could not find container \"0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115\": container with ID starting with 0721100c4019bedcb60c10a3c62147bcc1a46e82658e52837620ba5be1fe8115 not found: ID does not exist" Sep 30 14:33:22 crc kubenswrapper[4799]: I0930 14:33:22.035615 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-svk57"] Sep 30 14:33:22 crc kubenswrapper[4799]: I0930 14:33:22.039979 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-svk57"] Sep 30 14:33:22 crc kubenswrapper[4799]: I0930 14:33:22.511680 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03c9ee3f-722e-4db8-a8f5-dc359486e4f4" path="/var/lib/kubelet/pods/03c9ee3f-722e-4db8-a8f5-dc359486e4f4/volumes" Sep 30 14:33:23 crc kubenswrapper[4799]: I0930 14:33:23.002536 4799 generic.go:334] "Generic (PLEG): container finished" podID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerID="c41f17c423a3cf7a92057284fa0395cccc3dce54cb3f54d1ae6ca36c3f4d8e83" exitCode=0 Sep 30 14:33:23 crc kubenswrapper[4799]: I0930 14:33:23.002579 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" event={"ID":"90186c62-dccd-4a14-908c-6b6b5c1750ba","Type":"ContainerDied","Data":"c41f17c423a3cf7a92057284fa0395cccc3dce54cb3f54d1ae6ca36c3f4d8e83"} Sep 30 14:33:24 crc kubenswrapper[4799]: I0930 14:33:24.013349 4799 generic.go:334] "Generic (PLEG): container finished" podID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerID="a2d82a91d7c2ae85c51684c5deacd41f3ba03fdb033c433acf05b6cdfc5c5c00" exitCode=0 Sep 30 14:33:24 crc kubenswrapper[4799]: I0930 14:33:24.013402 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" event={"ID":"90186c62-dccd-4a14-908c-6b6b5c1750ba","Type":"ContainerDied","Data":"a2d82a91d7c2ae85c51684c5deacd41f3ba03fdb033c433acf05b6cdfc5c5c00"} Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.269102 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.391093 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9njk\" (UniqueName: \"kubernetes.io/projected/90186c62-dccd-4a14-908c-6b6b5c1750ba-kube-api-access-m9njk\") pod \"90186c62-dccd-4a14-908c-6b6b5c1750ba\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.391170 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-bundle\") pod \"90186c62-dccd-4a14-908c-6b6b5c1750ba\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.391209 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-util\") pod \"90186c62-dccd-4a14-908c-6b6b5c1750ba\" (UID: \"90186c62-dccd-4a14-908c-6b6b5c1750ba\") " Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.392328 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-bundle" (OuterVolumeSpecName: "bundle") pod "90186c62-dccd-4a14-908c-6b6b5c1750ba" (UID: "90186c62-dccd-4a14-908c-6b6b5c1750ba"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.399985 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90186c62-dccd-4a14-908c-6b6b5c1750ba-kube-api-access-m9njk" (OuterVolumeSpecName: "kube-api-access-m9njk") pod "90186c62-dccd-4a14-908c-6b6b5c1750ba" (UID: "90186c62-dccd-4a14-908c-6b6b5c1750ba"). InnerVolumeSpecName "kube-api-access-m9njk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.406370 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-util" (OuterVolumeSpecName: "util") pod "90186c62-dccd-4a14-908c-6b6b5c1750ba" (UID: "90186c62-dccd-4a14-908c-6b6b5c1750ba"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.492499 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9njk\" (UniqueName: \"kubernetes.io/projected/90186c62-dccd-4a14-908c-6b6b5c1750ba-kube-api-access-m9njk\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.492535 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:25 crc kubenswrapper[4799]: I0930 14:33:25.492543 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/90186c62-dccd-4a14-908c-6b6b5c1750ba-util\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:26 crc kubenswrapper[4799]: I0930 14:33:26.025331 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" event={"ID":"90186c62-dccd-4a14-908c-6b6b5c1750ba","Type":"ContainerDied","Data":"362e2ea46d11834aa1e6fa1667f386c9da1439eec21d89687360381199f93248"} Sep 30 14:33:26 crc kubenswrapper[4799]: I0930 14:33:26.025370 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="362e2ea46d11834aa1e6fa1667f386c9da1439eec21d89687360381199f93248" Sep 30 14:33:26 crc kubenswrapper[4799]: I0930 14:33:26.025371 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b" Sep 30 14:33:27 crc kubenswrapper[4799]: I0930 14:33:27.266341 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:27 crc kubenswrapper[4799]: I0930 14:33:27.267828 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:27 crc kubenswrapper[4799]: I0930 14:33:27.352809 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.090043 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.358739 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h29ms"] Sep 30 14:33:28 crc kubenswrapper[4799]: E0930 14:33:28.359375 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerName="util" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.359393 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerName="util" Sep 30 14:33:28 crc kubenswrapper[4799]: E0930 14:33:28.359412 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerName="pull" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.359423 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerName="pull" Sep 30 14:33:28 crc kubenswrapper[4799]: E0930 14:33:28.359433 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerName="extract" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.359440 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerName="extract" Sep 30 14:33:28 crc kubenswrapper[4799]: E0930 14:33:28.359453 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c9ee3f-722e-4db8-a8f5-dc359486e4f4" containerName="console" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.359460 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c9ee3f-722e-4db8-a8f5-dc359486e4f4" containerName="console" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.360315 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="90186c62-dccd-4a14-908c-6b6b5c1750ba" containerName="extract" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.360354 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="03c9ee3f-722e-4db8-a8f5-dc359486e4f4" containerName="console" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.371100 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.419313 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h29ms"] Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.529753 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kfhw\" (UniqueName: \"kubernetes.io/projected/23a721e7-44ad-40d9-8a3d-5b101240ed1e-kube-api-access-4kfhw\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.529825 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-utilities\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.529891 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-catalog-content\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.631237 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-catalog-content\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.631436 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kfhw\" (UniqueName: \"kubernetes.io/projected/23a721e7-44ad-40d9-8a3d-5b101240ed1e-kube-api-access-4kfhw\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.631485 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-utilities\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.632379 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-catalog-content\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.633463 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-utilities\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.664480 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kfhw\" (UniqueName: \"kubernetes.io/projected/23a721e7-44ad-40d9-8a3d-5b101240ed1e-kube-api-access-4kfhw\") pod \"certified-operators-h29ms\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:28 crc kubenswrapper[4799]: I0930 14:33:28.731256 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:29 crc kubenswrapper[4799]: I0930 14:33:29.163608 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h29ms"] Sep 30 14:33:30 crc kubenswrapper[4799]: I0930 14:33:30.067813 4799 generic.go:334] "Generic (PLEG): container finished" podID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerID="d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170" exitCode=0 Sep 30 14:33:30 crc kubenswrapper[4799]: I0930 14:33:30.068850 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h29ms" event={"ID":"23a721e7-44ad-40d9-8a3d-5b101240ed1e","Type":"ContainerDied","Data":"d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170"} Sep 30 14:33:30 crc kubenswrapper[4799]: I0930 14:33:30.068945 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h29ms" event={"ID":"23a721e7-44ad-40d9-8a3d-5b101240ed1e","Type":"ContainerStarted","Data":"847bf7c8ad711b30a1f5e23a19837fc36745c6ababa7a5b085061a90fb5ae942"} Sep 30 14:33:30 crc kubenswrapper[4799]: I0930 14:33:30.733657 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cr57c"] Sep 30 14:33:30 crc kubenswrapper[4799]: I0930 14:33:30.734217 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cr57c" podUID="185405f2-a875-49de-9107-ec4772cc69a5" containerName="registry-server" containerID="cri-o://a1f564328632914ce92ac09d826d628835491de2713fec5554c6e3a86e3e5401" gracePeriod=2 Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.108492 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h29ms" event={"ID":"23a721e7-44ad-40d9-8a3d-5b101240ed1e","Type":"ContainerStarted","Data":"b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c"} Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.113929 4799 generic.go:334] "Generic (PLEG): container finished" podID="185405f2-a875-49de-9107-ec4772cc69a5" containerID="a1f564328632914ce92ac09d826d628835491de2713fec5554c6e3a86e3e5401" exitCode=0 Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.113967 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr57c" event={"ID":"185405f2-a875-49de-9107-ec4772cc69a5","Type":"ContainerDied","Data":"a1f564328632914ce92ac09d826d628835491de2713fec5554c6e3a86e3e5401"} Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.234775 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.389265 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4jwd\" (UniqueName: \"kubernetes.io/projected/185405f2-a875-49de-9107-ec4772cc69a5-kube-api-access-n4jwd\") pod \"185405f2-a875-49de-9107-ec4772cc69a5\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.389323 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-utilities\") pod \"185405f2-a875-49de-9107-ec4772cc69a5\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.389391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-catalog-content\") pod \"185405f2-a875-49de-9107-ec4772cc69a5\" (UID: \"185405f2-a875-49de-9107-ec4772cc69a5\") " Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.390868 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-utilities" (OuterVolumeSpecName: "utilities") pod "185405f2-a875-49de-9107-ec4772cc69a5" (UID: "185405f2-a875-49de-9107-ec4772cc69a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.394252 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185405f2-a875-49de-9107-ec4772cc69a5-kube-api-access-n4jwd" (OuterVolumeSpecName: "kube-api-access-n4jwd") pod "185405f2-a875-49de-9107-ec4772cc69a5" (UID: "185405f2-a875-49de-9107-ec4772cc69a5"). InnerVolumeSpecName "kube-api-access-n4jwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.448481 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "185405f2-a875-49de-9107-ec4772cc69a5" (UID: "185405f2-a875-49de-9107-ec4772cc69a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.491084 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4jwd\" (UniqueName: \"kubernetes.io/projected/185405f2-a875-49de-9107-ec4772cc69a5-kube-api-access-n4jwd\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.491125 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:31 crc kubenswrapper[4799]: I0930 14:33:31.491136 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185405f2-a875-49de-9107-ec4772cc69a5-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.120852 4799 generic.go:334] "Generic (PLEG): container finished" podID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerID="b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c" exitCode=0 Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.120918 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h29ms" event={"ID":"23a721e7-44ad-40d9-8a3d-5b101240ed1e","Type":"ContainerDied","Data":"b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c"} Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.123067 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr57c" event={"ID":"185405f2-a875-49de-9107-ec4772cc69a5","Type":"ContainerDied","Data":"28470333e857e4ab6c3c983b1812f68a65ee6c14e1bc4c4b0ad6d3942d7acf7f"} Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.123115 4799 scope.go:117] "RemoveContainer" containerID="a1f564328632914ce92ac09d826d628835491de2713fec5554c6e3a86e3e5401" Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.123228 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cr57c" Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.148232 4799 scope.go:117] "RemoveContainer" containerID="0401d2f91c708e82578e9b1bb8d3fe561e6431d843d0d779de6dce29bd22f269" Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.195926 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cr57c"] Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.201008 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cr57c"] Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.201369 4799 scope.go:117] "RemoveContainer" containerID="f7486679dc40c58b850350509c501d6e3384f62f5908997248ebd44cab0a5779" Sep 30 14:33:32 crc kubenswrapper[4799]: I0930 14:33:32.511520 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185405f2-a875-49de-9107-ec4772cc69a5" path="/var/lib/kubelet/pods/185405f2-a875-49de-9107-ec4772cc69a5/volumes" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.130459 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h29ms" event={"ID":"23a721e7-44ad-40d9-8a3d-5b101240ed1e","Type":"ContainerStarted","Data":"14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a"} Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.175903 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h29ms" podStartSLOduration=2.656257355 podStartE2EDuration="5.175884815s" podCreationTimestamp="2025-09-30 14:33:28 +0000 UTC" firstStartedPulling="2025-09-30 14:33:30.071105061 +0000 UTC m=+832.154705478" lastFinishedPulling="2025-09-30 14:33:32.590732511 +0000 UTC m=+834.674332938" observedRunningTime="2025-09-30 14:33:33.173823725 +0000 UTC m=+835.257424172" watchObservedRunningTime="2025-09-30 14:33:33.175884815 +0000 UTC m=+835.259485242" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.530526 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh"] Sep 30 14:33:33 crc kubenswrapper[4799]: E0930 14:33:33.530842 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185405f2-a875-49de-9107-ec4772cc69a5" containerName="extract-content" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.530864 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="185405f2-a875-49de-9107-ec4772cc69a5" containerName="extract-content" Sep 30 14:33:33 crc kubenswrapper[4799]: E0930 14:33:33.530885 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185405f2-a875-49de-9107-ec4772cc69a5" containerName="extract-utilities" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.530895 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="185405f2-a875-49de-9107-ec4772cc69a5" containerName="extract-utilities" Sep 30 14:33:33 crc kubenswrapper[4799]: E0930 14:33:33.530917 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185405f2-a875-49de-9107-ec4772cc69a5" containerName="registry-server" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.530924 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="185405f2-a875-49de-9107-ec4772cc69a5" containerName="registry-server" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.531043 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="185405f2-a875-49de-9107-ec4772cc69a5" containerName="registry-server" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.531473 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.535079 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.535822 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fhj2k" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.536032 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.536171 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.536311 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.551548 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh"] Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.619471 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72f552b4-b1cb-4659-8629-d8ff13a85cc5-apiservice-cert\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.619551 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb8km\" (UniqueName: \"kubernetes.io/projected/72f552b4-b1cb-4659-8629-d8ff13a85cc5-kube-api-access-mb8km\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.619592 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72f552b4-b1cb-4659-8629-d8ff13a85cc5-webhook-cert\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.720570 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72f552b4-b1cb-4659-8629-d8ff13a85cc5-webhook-cert\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.720685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72f552b4-b1cb-4659-8629-d8ff13a85cc5-apiservice-cert\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.720720 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb8km\" (UniqueName: \"kubernetes.io/projected/72f552b4-b1cb-4659-8629-d8ff13a85cc5-kube-api-access-mb8km\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.726613 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72f552b4-b1cb-4659-8629-d8ff13a85cc5-apiservice-cert\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.737202 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72f552b4-b1cb-4659-8629-d8ff13a85cc5-webhook-cert\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.758399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb8km\" (UniqueName: \"kubernetes.io/projected/72f552b4-b1cb-4659-8629-d8ff13a85cc5-kube-api-access-mb8km\") pod \"metallb-operator-controller-manager-668b984c56-cwmvh\" (UID: \"72f552b4-b1cb-4659-8629-d8ff13a85cc5\") " pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.848937 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.870902 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j"] Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.872027 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.892412 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j"] Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.894271 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.894498 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Sep 30 14:33:33 crc kubenswrapper[4799]: I0930 14:33:33.894691 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-vddh8" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.034454 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7w8v\" (UniqueName: \"kubernetes.io/projected/624ab48f-5623-4bf1-af7f-4cad0be3e654-kube-api-access-p7w8v\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.034837 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/624ab48f-5623-4bf1-af7f-4cad0be3e654-apiservice-cert\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.034883 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/624ab48f-5623-4bf1-af7f-4cad0be3e654-webhook-cert\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.136575 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7w8v\" (UniqueName: \"kubernetes.io/projected/624ab48f-5623-4bf1-af7f-4cad0be3e654-kube-api-access-p7w8v\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.136687 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/624ab48f-5623-4bf1-af7f-4cad0be3e654-apiservice-cert\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.136719 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/624ab48f-5623-4bf1-af7f-4cad0be3e654-webhook-cert\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.149460 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/624ab48f-5623-4bf1-af7f-4cad0be3e654-apiservice-cert\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.149494 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/624ab48f-5623-4bf1-af7f-4cad0be3e654-webhook-cert\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.160706 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7w8v\" (UniqueName: \"kubernetes.io/projected/624ab48f-5623-4bf1-af7f-4cad0be3e654-kube-api-access-p7w8v\") pod \"metallb-operator-webhook-server-698f89f9d8-jdx5j\" (UID: \"624ab48f-5623-4bf1-af7f-4cad0be3e654\") " pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.251974 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.376247 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh"] Sep 30 14:33:34 crc kubenswrapper[4799]: W0930 14:33:34.402973 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72f552b4_b1cb_4659_8629_d8ff13a85cc5.slice/crio-c0ee3169b8ddae1c9c27789c6f391a6c74621287764edddbf2462cec91585318 WatchSource:0}: Error finding container c0ee3169b8ddae1c9c27789c6f391a6c74621287764edddbf2462cec91585318: Status 404 returned error can't find the container with id c0ee3169b8ddae1c9c27789c6f391a6c74621287764edddbf2462cec91585318 Sep 30 14:33:34 crc kubenswrapper[4799]: I0930 14:33:34.662453 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j"] Sep 30 14:33:34 crc kubenswrapper[4799]: W0930 14:33:34.672882 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod624ab48f_5623_4bf1_af7f_4cad0be3e654.slice/crio-59cb1649bdab2f9d40732bbb207fac9deae9fbfe9d92846a34da53f024de6d96 WatchSource:0}: Error finding container 59cb1649bdab2f9d40732bbb207fac9deae9fbfe9d92846a34da53f024de6d96: Status 404 returned error can't find the container with id 59cb1649bdab2f9d40732bbb207fac9deae9fbfe9d92846a34da53f024de6d96 Sep 30 14:33:35 crc kubenswrapper[4799]: I0930 14:33:35.158052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" event={"ID":"72f552b4-b1cb-4659-8629-d8ff13a85cc5","Type":"ContainerStarted","Data":"c0ee3169b8ddae1c9c27789c6f391a6c74621287764edddbf2462cec91585318"} Sep 30 14:33:35 crc kubenswrapper[4799]: I0930 14:33:35.159039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" event={"ID":"624ab48f-5623-4bf1-af7f-4cad0be3e654","Type":"ContainerStarted","Data":"59cb1649bdab2f9d40732bbb207fac9deae9fbfe9d92846a34da53f024de6d96"} Sep 30 14:33:38 crc kubenswrapper[4799]: I0930 14:33:38.732181 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:38 crc kubenswrapper[4799]: I0930 14:33:38.732557 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:38 crc kubenswrapper[4799]: I0930 14:33:38.796153 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:39 crc kubenswrapper[4799]: I0930 14:33:39.265187 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:39 crc kubenswrapper[4799]: I0930 14:33:39.936598 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h29ms"] Sep 30 14:33:41 crc kubenswrapper[4799]: I0930 14:33:41.207500 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h29ms" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerName="registry-server" containerID="cri-o://14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a" gracePeriod=2 Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.090035 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.213922 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" event={"ID":"72f552b4-b1cb-4659-8629-d8ff13a85cc5","Type":"ContainerStarted","Data":"af1ef66ea421ebce341efd4716e8021941b43ba408487268b037bc708d0704eb"} Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.214760 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.219939 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" event={"ID":"624ab48f-5623-4bf1-af7f-4cad0be3e654","Type":"ContainerStarted","Data":"14ba3e6fd2f64bf0aa371d2ba6d430ca660061b6d6c3a3afa8e754651504784c"} Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.219991 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.222216 4799 generic.go:334] "Generic (PLEG): container finished" podID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerID="14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a" exitCode=0 Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.222251 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h29ms" event={"ID":"23a721e7-44ad-40d9-8a3d-5b101240ed1e","Type":"ContainerDied","Data":"14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a"} Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.222271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h29ms" event={"ID":"23a721e7-44ad-40d9-8a3d-5b101240ed1e","Type":"ContainerDied","Data":"847bf7c8ad711b30a1f5e23a19837fc36745c6ababa7a5b085061a90fb5ae942"} Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.222290 4799 scope.go:117] "RemoveContainer" containerID="14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.222402 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h29ms" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.235968 4799 scope.go:117] "RemoveContainer" containerID="b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.243869 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" podStartSLOduration=1.844810791 podStartE2EDuration="9.243846768s" podCreationTimestamp="2025-09-30 14:33:33 +0000 UTC" firstStartedPulling="2025-09-30 14:33:34.411436558 +0000 UTC m=+836.495036995" lastFinishedPulling="2025-09-30 14:33:41.810472545 +0000 UTC m=+843.894072972" observedRunningTime="2025-09-30 14:33:42.241933502 +0000 UTC m=+844.325533939" watchObservedRunningTime="2025-09-30 14:33:42.243846768 +0000 UTC m=+844.327447195" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.259207 4799 scope.go:117] "RemoveContainer" containerID="d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.265891 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-catalog-content\") pod \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.265988 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kfhw\" (UniqueName: \"kubernetes.io/projected/23a721e7-44ad-40d9-8a3d-5b101240ed1e-kube-api-access-4kfhw\") pod \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.266055 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-utilities\") pod \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\" (UID: \"23a721e7-44ad-40d9-8a3d-5b101240ed1e\") " Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.267009 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-utilities" (OuterVolumeSpecName: "utilities") pod "23a721e7-44ad-40d9-8a3d-5b101240ed1e" (UID: "23a721e7-44ad-40d9-8a3d-5b101240ed1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.269469 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" podStartSLOduration=2.119384707 podStartE2EDuration="9.269449055s" podCreationTimestamp="2025-09-30 14:33:33 +0000 UTC" firstStartedPulling="2025-09-30 14:33:34.676208798 +0000 UTC m=+836.759809225" lastFinishedPulling="2025-09-30 14:33:41.826273146 +0000 UTC m=+843.909873573" observedRunningTime="2025-09-30 14:33:42.267374445 +0000 UTC m=+844.350974872" watchObservedRunningTime="2025-09-30 14:33:42.269449055 +0000 UTC m=+844.353049482" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.271272 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23a721e7-44ad-40d9-8a3d-5b101240ed1e-kube-api-access-4kfhw" (OuterVolumeSpecName: "kube-api-access-4kfhw") pod "23a721e7-44ad-40d9-8a3d-5b101240ed1e" (UID: "23a721e7-44ad-40d9-8a3d-5b101240ed1e"). InnerVolumeSpecName "kube-api-access-4kfhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.277535 4799 scope.go:117] "RemoveContainer" containerID="14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a" Sep 30 14:33:42 crc kubenswrapper[4799]: E0930 14:33:42.279225 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a\": container with ID starting with 14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a not found: ID does not exist" containerID="14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.279266 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a"} err="failed to get container status \"14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a\": rpc error: code = NotFound desc = could not find container \"14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a\": container with ID starting with 14a73232295930431c88eeeaad48a5a63853a57cc97dd15432626e86bd763d5a not found: ID does not exist" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.279292 4799 scope.go:117] "RemoveContainer" containerID="b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c" Sep 30 14:33:42 crc kubenswrapper[4799]: E0930 14:33:42.279782 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c\": container with ID starting with b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c not found: ID does not exist" containerID="b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.279806 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c"} err="failed to get container status \"b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c\": rpc error: code = NotFound desc = could not find container \"b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c\": container with ID starting with b165c88a1f3a22174da86da550b5232deba6e966b92101b139b6153b8241169c not found: ID does not exist" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.279825 4799 scope.go:117] "RemoveContainer" containerID="d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170" Sep 30 14:33:42 crc kubenswrapper[4799]: E0930 14:33:42.280046 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170\": container with ID starting with d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170 not found: ID does not exist" containerID="d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.280069 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170"} err="failed to get container status \"d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170\": rpc error: code = NotFound desc = could not find container \"d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170\": container with ID starting with d64468c6592158d18a627398cec4093d790b05ed95a919f7d54b97ec0eeac170 not found: ID does not exist" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.333791 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23a721e7-44ad-40d9-8a3d-5b101240ed1e" (UID: "23a721e7-44ad-40d9-8a3d-5b101240ed1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.368056 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.368928 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kfhw\" (UniqueName: \"kubernetes.io/projected/23a721e7-44ad-40d9-8a3d-5b101240ed1e-kube-api-access-4kfhw\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.368987 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a721e7-44ad-40d9-8a3d-5b101240ed1e-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.538977 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h29ms"] Sep 30 14:33:42 crc kubenswrapper[4799]: I0930 14:33:42.542004 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h29ms"] Sep 30 14:33:44 crc kubenswrapper[4799]: I0930 14:33:44.510948 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" path="/var/lib/kubelet/pods/23a721e7-44ad-40d9-8a3d-5b101240ed1e/volumes" Sep 30 14:33:54 crc kubenswrapper[4799]: I0930 14:33:54.258734 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-698f89f9d8-jdx5j" Sep 30 14:34:13 crc kubenswrapper[4799]: I0930 14:34:13.852499 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-668b984c56-cwmvh" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.676096 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r"] Sep 30 14:34:14 crc kubenswrapper[4799]: E0930 14:34:14.676731 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerName="extract-utilities" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.676755 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerName="extract-utilities" Sep 30 14:34:14 crc kubenswrapper[4799]: E0930 14:34:14.676769 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerName="extract-content" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.676777 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerName="extract-content" Sep 30 14:34:14 crc kubenswrapper[4799]: E0930 14:34:14.676788 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerName="registry-server" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.676799 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerName="registry-server" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.676923 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="23a721e7-44ad-40d9-8a3d-5b101240ed1e" containerName="registry-server" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.677436 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.686229 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ffzj6" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.686766 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.689202 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-dhdb5"] Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.694083 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.696133 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.696473 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.708466 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r"] Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821394 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl9f4\" (UniqueName: \"kubernetes.io/projected/6d15288f-3467-4f02-815d-ce489f8f5ad3-kube-api-access-sl9f4\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821459 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-sockets\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh6v9\" (UniqueName: \"kubernetes.io/projected/17444fcf-15e8-4038-9500-508d084a38cd-kube-api-access-lh6v9\") pod \"frr-k8s-webhook-server-5478bdb765-z8j6r\" (UID: \"17444fcf-15e8-4038-9500-508d084a38cd\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821511 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-conf\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821537 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics-certs\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821554 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821718 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-startup\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821760 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-reloader\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.821802 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17444fcf-15e8-4038-9500-508d084a38cd-cert\") pod \"frr-k8s-webhook-server-5478bdb765-z8j6r\" (UID: \"17444fcf-15e8-4038-9500-508d084a38cd\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.855930 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-mtwhr"] Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.856958 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mtwhr" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.864022 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.866903 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.867293 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.871010 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-s256w" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.918447 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-dcxwg"] Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.919592 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927422 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-startup\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927509 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-reloader\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927544 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17444fcf-15e8-4038-9500-508d084a38cd-cert\") pod \"frr-k8s-webhook-server-5478bdb765-z8j6r\" (UID: \"17444fcf-15e8-4038-9500-508d084a38cd\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927617 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl9f4\" (UniqueName: \"kubernetes.io/projected/6d15288f-3467-4f02-815d-ce489f8f5ad3-kube-api-access-sl9f4\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927677 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-sockets\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927712 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh6v9\" (UniqueName: \"kubernetes.io/projected/17444fcf-15e8-4038-9500-508d084a38cd-kube-api-access-lh6v9\") pod \"frr-k8s-webhook-server-5478bdb765-z8j6r\" (UID: \"17444fcf-15e8-4038-9500-508d084a38cd\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927736 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-conf\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927770 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics-certs\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.927793 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: E0930 14:34:14.928231 4799 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Sep 30 14:34:14 crc kubenswrapper[4799]: E0930 14:34:14.928303 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics-certs podName:6d15288f-3467-4f02-815d-ce489f8f5ad3 nodeName:}" failed. No retries permitted until 2025-09-30 14:34:15.428282326 +0000 UTC m=+877.511882833 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics-certs") pod "frr-k8s-dhdb5" (UID: "6d15288f-3467-4f02-815d-ce489f8f5ad3") : secret "frr-k8s-certs-secret" not found Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.928473 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-conf\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: E0930 14:34:14.928548 4799 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.928544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-reloader\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: E0930 14:34:14.928596 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/17444fcf-15e8-4038-9500-508d084a38cd-cert podName:17444fcf-15e8-4038-9500-508d084a38cd nodeName:}" failed. No retries permitted until 2025-09-30 14:34:15.428585244 +0000 UTC m=+877.512185761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/17444fcf-15e8-4038-9500-508d084a38cd-cert") pod "frr-k8s-webhook-server-5478bdb765-z8j6r" (UID: "17444fcf-15e8-4038-9500-508d084a38cd") : secret "frr-k8s-webhook-server-cert" not found Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.929113 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.929113 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-sockets\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.929335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.929427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6d15288f-3467-4f02-815d-ce489f8f5ad3-frr-startup\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.954768 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-dcxwg"] Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.966961 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl9f4\" (UniqueName: \"kubernetes.io/projected/6d15288f-3467-4f02-815d-ce489f8f5ad3-kube-api-access-sl9f4\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:14 crc kubenswrapper[4799]: I0930 14:34:14.985495 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh6v9\" (UniqueName: \"kubernetes.io/projected/17444fcf-15e8-4038-9500-508d084a38cd-kube-api-access-lh6v9\") pod \"frr-k8s-webhook-server-5478bdb765-z8j6r\" (UID: \"17444fcf-15e8-4038-9500-508d084a38cd\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.029293 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/056d4eef-c8d4-4f76-9006-cb6454ec953c-metallb-excludel2\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.029606 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/96370b88-b581-4342-a570-059733689d3e-cert\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.029763 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg2tx\" (UniqueName: \"kubernetes.io/projected/056d4eef-c8d4-4f76-9006-cb6454ec953c-kube-api-access-jg2tx\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.029886 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.030008 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fczrk\" (UniqueName: \"kubernetes.io/projected/96370b88-b581-4342-a570-059733689d3e-kube-api-access-fczrk\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.030127 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-metrics-certs\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.030419 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96370b88-b581-4342-a570-059733689d3e-metrics-certs\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.131812 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96370b88-b581-4342-a570-059733689d3e-metrics-certs\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.131867 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/056d4eef-c8d4-4f76-9006-cb6454ec953c-metallb-excludel2\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.131887 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/96370b88-b581-4342-a570-059733689d3e-cert\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.131913 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg2tx\" (UniqueName: \"kubernetes.io/projected/056d4eef-c8d4-4f76-9006-cb6454ec953c-kube-api-access-jg2tx\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.131940 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.131966 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fczrk\" (UniqueName: \"kubernetes.io/projected/96370b88-b581-4342-a570-059733689d3e-kube-api-access-fczrk\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.131990 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-metrics-certs\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: E0930 14:34:15.132141 4799 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Sep 30 14:34:15 crc kubenswrapper[4799]: E0930 14:34:15.132191 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-metrics-certs podName:056d4eef-c8d4-4f76-9006-cb6454ec953c nodeName:}" failed. No retries permitted until 2025-09-30 14:34:15.632176055 +0000 UTC m=+877.715776472 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-metrics-certs") pod "speaker-mtwhr" (UID: "056d4eef-c8d4-4f76-9006-cb6454ec953c") : secret "speaker-certs-secret" not found Sep 30 14:34:15 crc kubenswrapper[4799]: E0930 14:34:15.133251 4799 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.133285 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/056d4eef-c8d4-4f76-9006-cb6454ec953c-metallb-excludel2\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: E0930 14:34:15.133357 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist podName:056d4eef-c8d4-4f76-9006-cb6454ec953c nodeName:}" failed. No retries permitted until 2025-09-30 14:34:15.633330979 +0000 UTC m=+877.716931486 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist") pod "speaker-mtwhr" (UID: "056d4eef-c8d4-4f76-9006-cb6454ec953c") : secret "metallb-memberlist" not found Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.136762 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/96370b88-b581-4342-a570-059733689d3e-metrics-certs\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.138363 4799 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.146386 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/96370b88-b581-4342-a570-059733689d3e-cert\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.158412 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg2tx\" (UniqueName: \"kubernetes.io/projected/056d4eef-c8d4-4f76-9006-cb6454ec953c-kube-api-access-jg2tx\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.164773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fczrk\" (UniqueName: \"kubernetes.io/projected/96370b88-b581-4342-a570-059733689d3e-kube-api-access-fczrk\") pod \"controller-5d688f5ffc-dcxwg\" (UID: \"96370b88-b581-4342-a570-059733689d3e\") " pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.241057 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.436957 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17444fcf-15e8-4038-9500-508d084a38cd-cert\") pod \"frr-k8s-webhook-server-5478bdb765-z8j6r\" (UID: \"17444fcf-15e8-4038-9500-508d084a38cd\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.437093 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics-certs\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.440997 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d15288f-3467-4f02-815d-ce489f8f5ad3-metrics-certs\") pod \"frr-k8s-dhdb5\" (UID: \"6d15288f-3467-4f02-815d-ce489f8f5ad3\") " pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.441381 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17444fcf-15e8-4038-9500-508d084a38cd-cert\") pod \"frr-k8s-webhook-server-5478bdb765-z8j6r\" (UID: \"17444fcf-15e8-4038-9500-508d084a38cd\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.547309 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-dcxwg"] Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.595746 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.618873 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.642580 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.642636 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-metrics-certs\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: E0930 14:34:15.642833 4799 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 30 14:34:15 crc kubenswrapper[4799]: E0930 14:34:15.642932 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist podName:056d4eef-c8d4-4f76-9006-cb6454ec953c nodeName:}" failed. No retries permitted until 2025-09-30 14:34:16.642910196 +0000 UTC m=+878.726510683 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist") pod "speaker-mtwhr" (UID: "056d4eef-c8d4-4f76-9006-cb6454ec953c") : secret "metallb-memberlist" not found Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.646732 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-metrics-certs\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:15 crc kubenswrapper[4799]: I0930 14:34:15.837708 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r"] Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.429822 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" event={"ID":"17444fcf-15e8-4038-9500-508d084a38cd","Type":"ContainerStarted","Data":"3a2801b254302dcdf110a674fc9d6981c7a085a5b046903ded59a083b71eef16"} Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.432757 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-dcxwg" event={"ID":"96370b88-b581-4342-a570-059733689d3e","Type":"ContainerStarted","Data":"2c08441f0c38247fbacb4aea81a2132b142ee1cf14b46451fbe2f9631a9fff00"} Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.432807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-dcxwg" event={"ID":"96370b88-b581-4342-a570-059733689d3e","Type":"ContainerStarted","Data":"cefd0b0ea1aae2168525936a6211e1bab7404704765ffed578c52b06869f4183"} Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.432821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-dcxwg" event={"ID":"96370b88-b581-4342-a570-059733689d3e","Type":"ContainerStarted","Data":"dadbddf835271f9084bf3b6ca57b19da92a43fcaf6f6eec222f61d0005f7a46c"} Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.432875 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.434571 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerStarted","Data":"7ab45913f0d01507d74b6820bb4ae69978694b891f2005770f7221712b15d4d6"} Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.450636 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-dcxwg" podStartSLOduration=2.450613919 podStartE2EDuration="2.450613919s" podCreationTimestamp="2025-09-30 14:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:34:16.447959951 +0000 UTC m=+878.531560388" watchObservedRunningTime="2025-09-30 14:34:16.450613919 +0000 UTC m=+878.534214346" Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.654994 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.661289 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/056d4eef-c8d4-4f76-9006-cb6454ec953c-memberlist\") pod \"speaker-mtwhr\" (UID: \"056d4eef-c8d4-4f76-9006-cb6454ec953c\") " pod="metallb-system/speaker-mtwhr" Sep 30 14:34:16 crc kubenswrapper[4799]: I0930 14:34:16.672806 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mtwhr" Sep 30 14:34:16 crc kubenswrapper[4799]: W0930 14:34:16.693573 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod056d4eef_c8d4_4f76_9006_cb6454ec953c.slice/crio-f3f6347a2da59bdfc48c5a602f2d1bb9a73cfcdbe1ffb73c0ec6e6a0d21fad3c WatchSource:0}: Error finding container f3f6347a2da59bdfc48c5a602f2d1bb9a73cfcdbe1ffb73c0ec6e6a0d21fad3c: Status 404 returned error can't find the container with id f3f6347a2da59bdfc48c5a602f2d1bb9a73cfcdbe1ffb73c0ec6e6a0d21fad3c Sep 30 14:34:17 crc kubenswrapper[4799]: I0930 14:34:17.465762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mtwhr" event={"ID":"056d4eef-c8d4-4f76-9006-cb6454ec953c","Type":"ContainerStarted","Data":"586c0fb352a7bf1fa0b24ce66f0e849e30925f9332a02ecc96835a8b9ea6f804"} Sep 30 14:34:17 crc kubenswrapper[4799]: I0930 14:34:17.466132 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mtwhr" event={"ID":"056d4eef-c8d4-4f76-9006-cb6454ec953c","Type":"ContainerStarted","Data":"f3f6347a2da59bdfc48c5a602f2d1bb9a73cfcdbe1ffb73c0ec6e6a0d21fad3c"} Sep 30 14:34:18 crc kubenswrapper[4799]: I0930 14:34:18.480446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mtwhr" event={"ID":"056d4eef-c8d4-4f76-9006-cb6454ec953c","Type":"ContainerStarted","Data":"cc692f3097bb9f0b962a06b1a1a824695adf4d1acda19d6d7cf86fde95a703e8"} Sep 30 14:34:18 crc kubenswrapper[4799]: I0930 14:34:18.480855 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-mtwhr" Sep 30 14:34:18 crc kubenswrapper[4799]: I0930 14:34:18.520376 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-mtwhr" podStartSLOduration=4.520354116 podStartE2EDuration="4.520354116s" podCreationTimestamp="2025-09-30 14:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:34:18.512059354 +0000 UTC m=+880.595659801" watchObservedRunningTime="2025-09-30 14:34:18.520354116 +0000 UTC m=+880.603954553" Sep 30 14:34:25 crc kubenswrapper[4799]: I0930 14:34:25.246755 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-dcxwg" Sep 30 14:34:26 crc kubenswrapper[4799]: I0930 14:34:26.541181 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" event={"ID":"17444fcf-15e8-4038-9500-508d084a38cd","Type":"ContainerStarted","Data":"64747babcd37b4506acbfa12e7b6cbb63388cb403797716c8b352460e9963394"} Sep 30 14:34:26 crc kubenswrapper[4799]: I0930 14:34:26.541699 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:26 crc kubenswrapper[4799]: I0930 14:34:26.543388 4799 generic.go:334] "Generic (PLEG): container finished" podID="6d15288f-3467-4f02-815d-ce489f8f5ad3" containerID="84a32ac8b965aacbbbf2c5a27aa2482e59380fee9bd36f22dbfec29bbd67355f" exitCode=0 Sep 30 14:34:26 crc kubenswrapper[4799]: I0930 14:34:26.543426 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerDied","Data":"84a32ac8b965aacbbbf2c5a27aa2482e59380fee9bd36f22dbfec29bbd67355f"} Sep 30 14:34:26 crc kubenswrapper[4799]: I0930 14:34:26.587445 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" podStartSLOduration=2.551656563 podStartE2EDuration="12.587422687s" podCreationTimestamp="2025-09-30 14:34:14 +0000 UTC" firstStartedPulling="2025-09-30 14:34:15.856362926 +0000 UTC m=+877.939963353" lastFinishedPulling="2025-09-30 14:34:25.89212905 +0000 UTC m=+887.975729477" observedRunningTime="2025-09-30 14:34:26.565862546 +0000 UTC m=+888.649462973" watchObservedRunningTime="2025-09-30 14:34:26.587422687 +0000 UTC m=+888.671023114" Sep 30 14:34:26 crc kubenswrapper[4799]: I0930 14:34:26.679376 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-mtwhr" Sep 30 14:34:27 crc kubenswrapper[4799]: I0930 14:34:27.556311 4799 generic.go:334] "Generic (PLEG): container finished" podID="6d15288f-3467-4f02-815d-ce489f8f5ad3" containerID="f0c396e78b53b7d489d37ee4a0ec571179932a89427cddcfda6200b1f2cd5a4f" exitCode=0 Sep 30 14:34:27 crc kubenswrapper[4799]: I0930 14:34:27.556402 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerDied","Data":"f0c396e78b53b7d489d37ee4a0ec571179932a89427cddcfda6200b1f2cd5a4f"} Sep 30 14:34:28 crc kubenswrapper[4799]: I0930 14:34:28.567577 4799 generic.go:334] "Generic (PLEG): container finished" podID="6d15288f-3467-4f02-815d-ce489f8f5ad3" containerID="e072aa6801f4971b476716b58102691e433d14094ea55dc8c288be0c35fb6cbd" exitCode=0 Sep 30 14:34:28 crc kubenswrapper[4799]: I0930 14:34:28.567905 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerDied","Data":"e072aa6801f4971b476716b58102691e433d14094ea55dc8c288be0c35fb6cbd"} Sep 30 14:34:29 crc kubenswrapper[4799]: I0930 14:34:29.582207 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerStarted","Data":"68320a3193c5086b953302c906b41a6a639751427d9636ec71a0f8f1eb5342c6"} Sep 30 14:34:29 crc kubenswrapper[4799]: I0930 14:34:29.582484 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerStarted","Data":"89ba1ec6a8145b546cfc43654f4633cab56e59735c122ad55cdd22b03c6a6633"} Sep 30 14:34:29 crc kubenswrapper[4799]: I0930 14:34:29.582495 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerStarted","Data":"94a255692e37badb0c4001fb7d5b7896e81aeb531b407756ed1dfc98bb2f30af"} Sep 30 14:34:29 crc kubenswrapper[4799]: I0930 14:34:29.582503 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerStarted","Data":"269f1539342906d05ed60731c3f737460d83abbea055a144d9e2ad4c1c919bcb"} Sep 30 14:34:29 crc kubenswrapper[4799]: I0930 14:34:29.582511 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerStarted","Data":"195f949cabb30971cc55b57f230ac949d37bcff76da5a33082365a2916933684"} Sep 30 14:34:29 crc kubenswrapper[4799]: I0930 14:34:29.648721 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:34:29 crc kubenswrapper[4799]: I0930 14:34:29.648768 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.036455 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qlrfj"] Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.037320 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qlrfj" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.057165 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qlrfj"] Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.071896 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.072215 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-nkt6h" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.072299 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.072357 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlk2c\" (UniqueName: \"kubernetes.io/projected/91aa488b-8251-4427-ae46-23b015642c46-kube-api-access-rlk2c\") pod \"openstack-operator-index-qlrfj\" (UID: \"91aa488b-8251-4427-ae46-23b015642c46\") " pod="openstack-operators/openstack-operator-index-qlrfj" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.174272 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlk2c\" (UniqueName: \"kubernetes.io/projected/91aa488b-8251-4427-ae46-23b015642c46-kube-api-access-rlk2c\") pod \"openstack-operator-index-qlrfj\" (UID: \"91aa488b-8251-4427-ae46-23b015642c46\") " pod="openstack-operators/openstack-operator-index-qlrfj" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.204785 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlk2c\" (UniqueName: \"kubernetes.io/projected/91aa488b-8251-4427-ae46-23b015642c46-kube-api-access-rlk2c\") pod \"openstack-operator-index-qlrfj\" (UID: \"91aa488b-8251-4427-ae46-23b015642c46\") " pod="openstack-operators/openstack-operator-index-qlrfj" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.353330 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qlrfj" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.613399 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-dhdb5" event={"ID":"6d15288f-3467-4f02-815d-ce489f8f5ad3","Type":"ContainerStarted","Data":"5d2ddaaa9b94907500b684e590d4d07fc83adccd0badf5bfc9a7d1f1d6175528"} Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.615388 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.620353 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.643275 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-dhdb5" podStartSLOduration=6.534070784 podStartE2EDuration="16.643253314s" podCreationTimestamp="2025-09-30 14:34:14 +0000 UTC" firstStartedPulling="2025-09-30 14:34:15.773553926 +0000 UTC m=+877.857154353" lastFinishedPulling="2025-09-30 14:34:25.882736456 +0000 UTC m=+887.966336883" observedRunningTime="2025-09-30 14:34:30.641300957 +0000 UTC m=+892.724901394" watchObservedRunningTime="2025-09-30 14:34:30.643253314 +0000 UTC m=+892.726853741" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.700069 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:30 crc kubenswrapper[4799]: I0930 14:34:30.911283 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qlrfj"] Sep 30 14:34:30 crc kubenswrapper[4799]: W0930 14:34:30.922338 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91aa488b_8251_4427_ae46_23b015642c46.slice/crio-991b6bc462dd0358decd48e51072329984c225123ce65b5201d9fcdbf1250214 WatchSource:0}: Error finding container 991b6bc462dd0358decd48e51072329984c225123ce65b5201d9fcdbf1250214: Status 404 returned error can't find the container with id 991b6bc462dd0358decd48e51072329984c225123ce65b5201d9fcdbf1250214 Sep 30 14:34:31 crc kubenswrapper[4799]: I0930 14:34:31.621121 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qlrfj" event={"ID":"91aa488b-8251-4427-ae46-23b015642c46","Type":"ContainerStarted","Data":"991b6bc462dd0358decd48e51072329984c225123ce65b5201d9fcdbf1250214"} Sep 30 14:34:33 crc kubenswrapper[4799]: I0930 14:34:33.415079 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qlrfj"] Sep 30 14:34:34 crc kubenswrapper[4799]: I0930 14:34:34.019677 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-w6tbf"] Sep 30 14:34:34 crc kubenswrapper[4799]: I0930 14:34:34.020733 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:34 crc kubenswrapper[4799]: I0930 14:34:34.030965 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-w6tbf"] Sep 30 14:34:34 crc kubenswrapper[4799]: I0930 14:34:34.141594 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z554n\" (UniqueName: \"kubernetes.io/projected/f2bdc984-2b22-4a12-99b0-ef6235691241-kube-api-access-z554n\") pod \"openstack-operator-index-w6tbf\" (UID: \"f2bdc984-2b22-4a12-99b0-ef6235691241\") " pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:34 crc kubenswrapper[4799]: I0930 14:34:34.242586 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z554n\" (UniqueName: \"kubernetes.io/projected/f2bdc984-2b22-4a12-99b0-ef6235691241-kube-api-access-z554n\") pod \"openstack-operator-index-w6tbf\" (UID: \"f2bdc984-2b22-4a12-99b0-ef6235691241\") " pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:34 crc kubenswrapper[4799]: I0930 14:34:34.278943 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z554n\" (UniqueName: \"kubernetes.io/projected/f2bdc984-2b22-4a12-99b0-ef6235691241-kube-api-access-z554n\") pod \"openstack-operator-index-w6tbf\" (UID: \"f2bdc984-2b22-4a12-99b0-ef6235691241\") " pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:34 crc kubenswrapper[4799]: I0930 14:34:34.350563 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:36 crc kubenswrapper[4799]: I0930 14:34:36.367027 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-w6tbf"] Sep 30 14:34:38 crc kubenswrapper[4799]: I0930 14:34:38.672179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w6tbf" event={"ID":"f2bdc984-2b22-4a12-99b0-ef6235691241","Type":"ContainerStarted","Data":"831257c6a6712f04ca0415770a87e9027e209ba38bc4b3806466026f9d2ec8bd"} Sep 30 14:34:40 crc kubenswrapper[4799]: I0930 14:34:40.689689 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-qlrfj" podUID="91aa488b-8251-4427-ae46-23b015642c46" containerName="registry-server" containerID="cri-o://c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d" gracePeriod=2 Sep 30 14:34:40 crc kubenswrapper[4799]: I0930 14:34:40.689755 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qlrfj" event={"ID":"91aa488b-8251-4427-ae46-23b015642c46","Type":"ContainerStarted","Data":"c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d"} Sep 30 14:34:40 crc kubenswrapper[4799]: I0930 14:34:40.698504 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w6tbf" event={"ID":"f2bdc984-2b22-4a12-99b0-ef6235691241","Type":"ContainerStarted","Data":"cb21e748a90dc5754d07100f18a5206522838b5338e4ee08759666a0700cf665"} Sep 30 14:34:40 crc kubenswrapper[4799]: I0930 14:34:40.735214 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qlrfj" podStartSLOduration=1.889593289 podStartE2EDuration="10.735172511s" podCreationTimestamp="2025-09-30 14:34:30 +0000 UTC" firstStartedPulling="2025-09-30 14:34:30.922403035 +0000 UTC m=+893.006003462" lastFinishedPulling="2025-09-30 14:34:39.767982257 +0000 UTC m=+901.851582684" observedRunningTime="2025-09-30 14:34:40.71770039 +0000 UTC m=+902.801300837" watchObservedRunningTime="2025-09-30 14:34:40.735172511 +0000 UTC m=+902.818772938" Sep 30 14:34:40 crc kubenswrapper[4799]: I0930 14:34:40.736745 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-w6tbf" podStartSLOduration=5.596770211 podStartE2EDuration="6.736729867s" podCreationTimestamp="2025-09-30 14:34:34 +0000 UTC" firstStartedPulling="2025-09-30 14:34:38.627708912 +0000 UTC m=+900.711309339" lastFinishedPulling="2025-09-30 14:34:39.767668568 +0000 UTC m=+901.851268995" observedRunningTime="2025-09-30 14:34:40.732561445 +0000 UTC m=+902.816161882" watchObservedRunningTime="2025-09-30 14:34:40.736729867 +0000 UTC m=+902.820330294" Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.061868 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qlrfj" Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.138663 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlk2c\" (UniqueName: \"kubernetes.io/projected/91aa488b-8251-4427-ae46-23b015642c46-kube-api-access-rlk2c\") pod \"91aa488b-8251-4427-ae46-23b015642c46\" (UID: \"91aa488b-8251-4427-ae46-23b015642c46\") " Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.150782 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91aa488b-8251-4427-ae46-23b015642c46-kube-api-access-rlk2c" (OuterVolumeSpecName: "kube-api-access-rlk2c") pod "91aa488b-8251-4427-ae46-23b015642c46" (UID: "91aa488b-8251-4427-ae46-23b015642c46"). InnerVolumeSpecName "kube-api-access-rlk2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.240162 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlk2c\" (UniqueName: \"kubernetes.io/projected/91aa488b-8251-4427-ae46-23b015642c46-kube-api-access-rlk2c\") on node \"crc\" DevicePath \"\"" Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.706771 4799 generic.go:334] "Generic (PLEG): container finished" podID="91aa488b-8251-4427-ae46-23b015642c46" containerID="c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d" exitCode=0 Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.706833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qlrfj" event={"ID":"91aa488b-8251-4427-ae46-23b015642c46","Type":"ContainerDied","Data":"c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d"} Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.707915 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qlrfj" event={"ID":"91aa488b-8251-4427-ae46-23b015642c46","Type":"ContainerDied","Data":"991b6bc462dd0358decd48e51072329984c225123ce65b5201d9fcdbf1250214"} Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.707949 4799 scope.go:117] "RemoveContainer" containerID="c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d" Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.706845 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qlrfj" Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.728552 4799 scope.go:117] "RemoveContainer" containerID="c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d" Sep 30 14:34:41 crc kubenswrapper[4799]: E0930 14:34:41.729048 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d\": container with ID starting with c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d not found: ID does not exist" containerID="c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d" Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.729096 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d"} err="failed to get container status \"c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d\": rpc error: code = NotFound desc = could not find container \"c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d\": container with ID starting with c4a9cdb3986b1332640ba8da317a6c23225a771589d2cc727453524059afbe5d not found: ID does not exist" Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.741145 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qlrfj"] Sep 30 14:34:41 crc kubenswrapper[4799]: I0930 14:34:41.748444 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-qlrfj"] Sep 30 14:34:42 crc kubenswrapper[4799]: I0930 14:34:42.511099 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91aa488b-8251-4427-ae46-23b015642c46" path="/var/lib/kubelet/pods/91aa488b-8251-4427-ae46-23b015642c46/volumes" Sep 30 14:34:44 crc kubenswrapper[4799]: I0930 14:34:44.351661 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:44 crc kubenswrapper[4799]: I0930 14:34:44.352054 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:44 crc kubenswrapper[4799]: I0930 14:34:44.385990 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:44 crc kubenswrapper[4799]: I0930 14:34:44.750157 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-w6tbf" Sep 30 14:34:45 crc kubenswrapper[4799]: I0930 14:34:45.605890 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-z8j6r" Sep 30 14:34:45 crc kubenswrapper[4799]: I0930 14:34:45.626343 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-dhdb5" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.271597 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd"] Sep 30 14:34:51 crc kubenswrapper[4799]: E0930 14:34:51.272553 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91aa488b-8251-4427-ae46-23b015642c46" containerName="registry-server" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.272572 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="91aa488b-8251-4427-ae46-23b015642c46" containerName="registry-server" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.272764 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="91aa488b-8251-4427-ae46-23b015642c46" containerName="registry-server" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.273827 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.276679 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-469t9" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.340169 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd"] Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.389462 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-bundle\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.389541 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-util\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.389593 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c46k8\" (UniqueName: \"kubernetes.io/projected/49fbf469-452c-400c-ba9d-8970fbe1abce-kube-api-access-c46k8\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.491301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-bundle\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.491393 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-util\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.491437 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c46k8\" (UniqueName: \"kubernetes.io/projected/49fbf469-452c-400c-ba9d-8970fbe1abce-kube-api-access-c46k8\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.492530 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-bundle\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.492850 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-util\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.516373 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c46k8\" (UniqueName: \"kubernetes.io/projected/49fbf469-452c-400c-ba9d-8970fbe1abce-kube-api-access-c46k8\") pod \"897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:51 crc kubenswrapper[4799]: I0930 14:34:51.590263 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:52 crc kubenswrapper[4799]: I0930 14:34:52.089616 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd"] Sep 30 14:34:52 crc kubenswrapper[4799]: W0930 14:34:52.095578 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49fbf469_452c_400c_ba9d_8970fbe1abce.slice/crio-2cc2619d16b04ea3175b86d63edbb0c7ce87928aa0493f52a126e600e337cd3d WatchSource:0}: Error finding container 2cc2619d16b04ea3175b86d63edbb0c7ce87928aa0493f52a126e600e337cd3d: Status 404 returned error can't find the container with id 2cc2619d16b04ea3175b86d63edbb0c7ce87928aa0493f52a126e600e337cd3d Sep 30 14:34:52 crc kubenswrapper[4799]: I0930 14:34:52.774786 4799 generic.go:334] "Generic (PLEG): container finished" podID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerID="07ed02c6dd5f99239ace9f35058804a249bc854e7f0d0d4a2b30cf95b461f3a8" exitCode=0 Sep 30 14:34:52 crc kubenswrapper[4799]: I0930 14:34:52.774919 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" event={"ID":"49fbf469-452c-400c-ba9d-8970fbe1abce","Type":"ContainerDied","Data":"07ed02c6dd5f99239ace9f35058804a249bc854e7f0d0d4a2b30cf95b461f3a8"} Sep 30 14:34:52 crc kubenswrapper[4799]: I0930 14:34:52.775899 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" event={"ID":"49fbf469-452c-400c-ba9d-8970fbe1abce","Type":"ContainerStarted","Data":"2cc2619d16b04ea3175b86d63edbb0c7ce87928aa0493f52a126e600e337cd3d"} Sep 30 14:34:53 crc kubenswrapper[4799]: I0930 14:34:53.784959 4799 generic.go:334] "Generic (PLEG): container finished" podID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerID="f88271345de88bf725f83242529b99bb12263a25fdf9f8ab58828ad79ac5ec5d" exitCode=0 Sep 30 14:34:53 crc kubenswrapper[4799]: I0930 14:34:53.785012 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" event={"ID":"49fbf469-452c-400c-ba9d-8970fbe1abce","Type":"ContainerDied","Data":"f88271345de88bf725f83242529b99bb12263a25fdf9f8ab58828ad79ac5ec5d"} Sep 30 14:34:54 crc kubenswrapper[4799]: I0930 14:34:54.794088 4799 generic.go:334] "Generic (PLEG): container finished" podID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerID="e893d483d25d48a1484ff894922c9c3adbd6765c24c953529a58d6bb016eff28" exitCode=0 Sep 30 14:34:54 crc kubenswrapper[4799]: I0930 14:34:54.794154 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" event={"ID":"49fbf469-452c-400c-ba9d-8970fbe1abce","Type":"ContainerDied","Data":"e893d483d25d48a1484ff894922c9c3adbd6765c24c953529a58d6bb016eff28"} Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.066536 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.160640 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-bundle\") pod \"49fbf469-452c-400c-ba9d-8970fbe1abce\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.160970 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-util\") pod \"49fbf469-452c-400c-ba9d-8970fbe1abce\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.162408 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-bundle" (OuterVolumeSpecName: "bundle") pod "49fbf469-452c-400c-ba9d-8970fbe1abce" (UID: "49fbf469-452c-400c-ba9d-8970fbe1abce"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.164473 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c46k8\" (UniqueName: \"kubernetes.io/projected/49fbf469-452c-400c-ba9d-8970fbe1abce-kube-api-access-c46k8\") pod \"49fbf469-452c-400c-ba9d-8970fbe1abce\" (UID: \"49fbf469-452c-400c-ba9d-8970fbe1abce\") " Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.165467 4799 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.170068 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49fbf469-452c-400c-ba9d-8970fbe1abce-kube-api-access-c46k8" (OuterVolumeSpecName: "kube-api-access-c46k8") pod "49fbf469-452c-400c-ba9d-8970fbe1abce" (UID: "49fbf469-452c-400c-ba9d-8970fbe1abce"). InnerVolumeSpecName "kube-api-access-c46k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.176596 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-util" (OuterVolumeSpecName: "util") pod "49fbf469-452c-400c-ba9d-8970fbe1abce" (UID: "49fbf469-452c-400c-ba9d-8970fbe1abce"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.266400 4799 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/49fbf469-452c-400c-ba9d-8970fbe1abce-util\") on node \"crc\" DevicePath \"\"" Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.266599 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c46k8\" (UniqueName: \"kubernetes.io/projected/49fbf469-452c-400c-ba9d-8970fbe1abce-kube-api-access-c46k8\") on node \"crc\" DevicePath \"\"" Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.810625 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" event={"ID":"49fbf469-452c-400c-ba9d-8970fbe1abce","Type":"ContainerDied","Data":"2cc2619d16b04ea3175b86d63edbb0c7ce87928aa0493f52a126e600e337cd3d"} Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.810694 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cc2619d16b04ea3175b86d63edbb0c7ce87928aa0493f52a126e600e337cd3d" Sep 30 14:34:56 crc kubenswrapper[4799]: I0930 14:34:56.810668 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd" Sep 30 14:34:59 crc kubenswrapper[4799]: I0930 14:34:59.649612 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:34:59 crc kubenswrapper[4799]: I0930 14:34:59.649984 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.846292 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q"] Sep 30 14:35:03 crc kubenswrapper[4799]: E0930 14:35:03.846900 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerName="util" Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.846919 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerName="util" Sep 30 14:35:03 crc kubenswrapper[4799]: E0930 14:35:03.846931 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerName="extract" Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.846938 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerName="extract" Sep 30 14:35:03 crc kubenswrapper[4799]: E0930 14:35:03.846958 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerName="pull" Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.846967 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerName="pull" Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.847091 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="49fbf469-452c-400c-ba9d-8970fbe1abce" containerName="extract" Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.847988 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.851629 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-9ghrx" Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.934672 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q"] Sep 30 14:35:03 crc kubenswrapper[4799]: I0930 14:35:03.968186 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2ts7\" (UniqueName: \"kubernetes.io/projected/2cfc3242-481e-4575-9d3a-bbd740b097f6-kube-api-access-l2ts7\") pod \"openstack-operator-controller-operator-655f477c67-z5l5q\" (UID: \"2cfc3242-481e-4575-9d3a-bbd740b097f6\") " pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" Sep 30 14:35:04 crc kubenswrapper[4799]: I0930 14:35:04.070252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2ts7\" (UniqueName: \"kubernetes.io/projected/2cfc3242-481e-4575-9d3a-bbd740b097f6-kube-api-access-l2ts7\") pod \"openstack-operator-controller-operator-655f477c67-z5l5q\" (UID: \"2cfc3242-481e-4575-9d3a-bbd740b097f6\") " pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" Sep 30 14:35:04 crc kubenswrapper[4799]: I0930 14:35:04.091963 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2ts7\" (UniqueName: \"kubernetes.io/projected/2cfc3242-481e-4575-9d3a-bbd740b097f6-kube-api-access-l2ts7\") pod \"openstack-operator-controller-operator-655f477c67-z5l5q\" (UID: \"2cfc3242-481e-4575-9d3a-bbd740b097f6\") " pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" Sep 30 14:35:04 crc kubenswrapper[4799]: I0930 14:35:04.171951 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" Sep 30 14:35:04 crc kubenswrapper[4799]: I0930 14:35:04.444759 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q"] Sep 30 14:35:04 crc kubenswrapper[4799]: I0930 14:35:04.860398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" event={"ID":"2cfc3242-481e-4575-9d3a-bbd740b097f6","Type":"ContainerStarted","Data":"033e5afa56b4c3c4f6ffba35b6f82969115b93ed48ad6f43bd0491280c462062"} Sep 30 14:35:10 crc kubenswrapper[4799]: I0930 14:35:10.919022 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" event={"ID":"2cfc3242-481e-4575-9d3a-bbd740b097f6","Type":"ContainerStarted","Data":"a906391b906e58396bd345e4ec8541110c3a44019b2d78dce2a3bd31c4b630a9"} Sep 30 14:35:12 crc kubenswrapper[4799]: I0930 14:35:12.935102 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" event={"ID":"2cfc3242-481e-4575-9d3a-bbd740b097f6","Type":"ContainerStarted","Data":"3cf5e1726925b07f9b9d6b6712c68253a6c23b387242200d310f8031f69bb332"} Sep 30 14:35:12 crc kubenswrapper[4799]: I0930 14:35:12.935503 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" Sep 30 14:35:12 crc kubenswrapper[4799]: I0930 14:35:12.967046 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" podStartSLOduration=1.701367062 podStartE2EDuration="9.967027029s" podCreationTimestamp="2025-09-30 14:35:03 +0000 UTC" firstStartedPulling="2025-09-30 14:35:04.459189952 +0000 UTC m=+926.542790379" lastFinishedPulling="2025-09-30 14:35:12.724849909 +0000 UTC m=+934.808450346" observedRunningTime="2025-09-30 14:35:12.96536567 +0000 UTC m=+935.048966107" watchObservedRunningTime="2025-09-30 14:35:12.967027029 +0000 UTC m=+935.050627456" Sep 30 14:35:24 crc kubenswrapper[4799]: I0930 14:35:24.175582 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-655f477c67-z5l5q" Sep 30 14:35:29 crc kubenswrapper[4799]: I0930 14:35:29.648894 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:35:29 crc kubenswrapper[4799]: I0930 14:35:29.649258 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:35:29 crc kubenswrapper[4799]: I0930 14:35:29.649317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:35:29 crc kubenswrapper[4799]: I0930 14:35:29.650059 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7be6e12e797fed54fc7aab60eaeaf0c5516974a723cfc906b94db6e0a39463b8"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:35:29 crc kubenswrapper[4799]: I0930 14:35:29.650286 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://7be6e12e797fed54fc7aab60eaeaf0c5516974a723cfc906b94db6e0a39463b8" gracePeriod=600 Sep 30 14:35:30 crc kubenswrapper[4799]: I0930 14:35:30.053190 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="7be6e12e797fed54fc7aab60eaeaf0c5516974a723cfc906b94db6e0a39463b8" exitCode=0 Sep 30 14:35:30 crc kubenswrapper[4799]: I0930 14:35:30.053245 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"7be6e12e797fed54fc7aab60eaeaf0c5516974a723cfc906b94db6e0a39463b8"} Sep 30 14:35:30 crc kubenswrapper[4799]: I0930 14:35:30.053281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"c08cc75dc1edb21f93545a035f6f554f287534552d52a3400113832177d88167"} Sep 30 14:35:30 crc kubenswrapper[4799]: I0930 14:35:30.053300 4799 scope.go:117] "RemoveContainer" containerID="1e25c11d4641ce442c772191497cb466d8f447d176e97bcb309720e08dd67a66" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.806397 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.809834 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.811822 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-bgz6j" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.811887 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.812650 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.814928 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-87hql" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.829585 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.848228 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.849497 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.850878 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-tndp6" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.868942 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.876784 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.906420 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.908663 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.915443 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-vqrts" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.945704 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.952531 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr9hs\" (UniqueName: \"kubernetes.io/projected/8bc24269-6185-41a5-861b-19c88576c223-kube-api-access-jr9hs\") pod \"designate-operator-controller-manager-84f4f7b77b-mplqn\" (UID: \"8bc24269-6185-41a5-861b-19c88576c223\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.952624 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfdpn\" (UniqueName: \"kubernetes.io/projected/e0002170-fb28-4c87-a970-350f92c891bd-kube-api-access-tfdpn\") pod \"cinder-operator-controller-manager-644bddb6d8-kqfcp\" (UID: \"e0002170-fb28-4c87-a970-350f92c891bd\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.952697 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2qb5\" (UniqueName: \"kubernetes.io/projected/b8a734fd-7d92-4b2d-9bb4-91eee56047c5-kube-api-access-f2qb5\") pod \"barbican-operator-controller-manager-6ff8b75857-j7tgf\" (UID: \"b8a734fd-7d92-4b2d-9bb4-91eee56047c5\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.952738 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9jsv\" (UniqueName: \"kubernetes.io/projected/19539cfa-6f83-4636-8193-d2373038f353-kube-api-access-r9jsv\") pod \"glance-operator-controller-manager-84958c4d49-gmxsp\" (UID: \"19539cfa-6f83-4636-8193-d2373038f353\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.962687 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2"] Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.964761 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" Sep 30 14:35:49 crc kubenswrapper[4799]: I0930 14:35:49.969139 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-pmh8b" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.003293 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.019793 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.020926 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.025001 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-99qmh" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.030747 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.047040 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.048641 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.055122 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-fsj9k" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.055361 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.055517 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9jsv\" (UniqueName: \"kubernetes.io/projected/19539cfa-6f83-4636-8193-d2373038f353-kube-api-access-r9jsv\") pod \"glance-operator-controller-manager-84958c4d49-gmxsp\" (UID: \"19539cfa-6f83-4636-8193-d2373038f353\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.056053 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr9hs\" (UniqueName: \"kubernetes.io/projected/8bc24269-6185-41a5-861b-19c88576c223-kube-api-access-jr9hs\") pod \"designate-operator-controller-manager-84f4f7b77b-mplqn\" (UID: \"8bc24269-6185-41a5-861b-19c88576c223\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.056136 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmqkr\" (UniqueName: \"kubernetes.io/projected/08704fb7-1721-4f10-97a1-f2359d01cdaf-kube-api-access-zmqkr\") pod \"heat-operator-controller-manager-5d889d78cf-8vxq2\" (UID: \"08704fb7-1721-4f10-97a1-f2359d01cdaf\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.056173 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfdpn\" (UniqueName: \"kubernetes.io/projected/e0002170-fb28-4c87-a970-350f92c891bd-kube-api-access-tfdpn\") pod \"cinder-operator-controller-manager-644bddb6d8-kqfcp\" (UID: \"e0002170-fb28-4c87-a970-350f92c891bd\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.056243 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2qb5\" (UniqueName: \"kubernetes.io/projected/b8a734fd-7d92-4b2d-9bb4-91eee56047c5-kube-api-access-f2qb5\") pod \"barbican-operator-controller-manager-6ff8b75857-j7tgf\" (UID: \"b8a734fd-7d92-4b2d-9bb4-91eee56047c5\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.089084 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.090169 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.097425 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-xqqpb" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.099981 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.113964 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.126022 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr9hs\" (UniqueName: \"kubernetes.io/projected/8bc24269-6185-41a5-861b-19c88576c223-kube-api-access-jr9hs\") pod \"designate-operator-controller-manager-84f4f7b77b-mplqn\" (UID: \"8bc24269-6185-41a5-861b-19c88576c223\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.128777 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.136931 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.140522 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9jsv\" (UniqueName: \"kubernetes.io/projected/19539cfa-6f83-4636-8193-d2373038f353-kube-api-access-r9jsv\") pod \"glance-operator-controller-manager-84958c4d49-gmxsp\" (UID: \"19539cfa-6f83-4636-8193-d2373038f353\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.148951 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-5bfwm" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.149157 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfdpn\" (UniqueName: \"kubernetes.io/projected/e0002170-fb28-4c87-a970-350f92c891bd-kube-api-access-tfdpn\") pod \"cinder-operator-controller-manager-644bddb6d8-kqfcp\" (UID: \"e0002170-fb28-4c87-a970-350f92c891bd\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.159519 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txpxg\" (UniqueName: \"kubernetes.io/projected/2e07068f-3f60-4b35-9465-1adb7cda6525-kube-api-access-txpxg\") pod \"ironic-operator-controller-manager-7975b88857-p75vq\" (UID: \"2e07068f-3f60-4b35-9465-1adb7cda6525\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.159584 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert\") pod \"infra-operator-controller-manager-7d857cc749-rzjpm\" (UID: \"cdca0d27-1d59-4858-85a4-681c1d6abf90\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.159681 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmqkr\" (UniqueName: \"kubernetes.io/projected/08704fb7-1721-4f10-97a1-f2359d01cdaf-kube-api-access-zmqkr\") pod \"heat-operator-controller-manager-5d889d78cf-8vxq2\" (UID: \"08704fb7-1721-4f10-97a1-f2359d01cdaf\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.159762 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv2cx\" (UniqueName: \"kubernetes.io/projected/fdb1977f-1341-46ae-a4db-e78e20b6d5bd-kube-api-access-rv2cx\") pod \"horizon-operator-controller-manager-9f4696d94-k955x\" (UID: \"fdb1977f-1341-46ae-a4db-e78e20b6d5bd\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.159790 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5xzr\" (UniqueName: \"kubernetes.io/projected/cdca0d27-1d59-4858-85a4-681c1d6abf90-kube-api-access-p5xzr\") pod \"infra-operator-controller-manager-7d857cc749-rzjpm\" (UID: \"cdca0d27-1d59-4858-85a4-681c1d6abf90\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.160280 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.163009 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2qb5\" (UniqueName: \"kubernetes.io/projected/b8a734fd-7d92-4b2d-9bb4-91eee56047c5-kube-api-access-f2qb5\") pod \"barbican-operator-controller-manager-6ff8b75857-j7tgf\" (UID: \"b8a734fd-7d92-4b2d-9bb4-91eee56047c5\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.169591 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.174239 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.175780 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.177863 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-9hxn8" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.199430 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.205571 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.206731 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.214512 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.217580 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-d76h8" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.242244 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.243809 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmqkr\" (UniqueName: \"kubernetes.io/projected/08704fb7-1721-4f10-97a1-f2359d01cdaf-kube-api-access-zmqkr\") pod \"heat-operator-controller-manager-5d889d78cf-8vxq2\" (UID: \"08704fb7-1721-4f10-97a1-f2359d01cdaf\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.261083 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv2cx\" (UniqueName: \"kubernetes.io/projected/fdb1977f-1341-46ae-a4db-e78e20b6d5bd-kube-api-access-rv2cx\") pod \"horizon-operator-controller-manager-9f4696d94-k955x\" (UID: \"fdb1977f-1341-46ae-a4db-e78e20b6d5bd\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.261129 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5xzr\" (UniqueName: \"kubernetes.io/projected/cdca0d27-1d59-4858-85a4-681c1d6abf90-kube-api-access-p5xzr\") pod \"infra-operator-controller-manager-7d857cc749-rzjpm\" (UID: \"cdca0d27-1d59-4858-85a4-681c1d6abf90\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.261159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk9qc\" (UniqueName: \"kubernetes.io/projected/c8d8a2c1-6828-49d3-bd9a-baa3fe58790e-kube-api-access-xk9qc\") pod \"mariadb-operator-controller-manager-88c7-pxd4j\" (UID: \"c8d8a2c1-6828-49d3-bd9a-baa3fe58790e\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.261184 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgnn7\" (UniqueName: \"kubernetes.io/projected/40600429-6eda-47b1-bf61-0e22189f7549-kube-api-access-pgnn7\") pod \"manila-operator-controller-manager-6d68dbc695-2jnc8\" (UID: \"40600429-6eda-47b1-bf61-0e22189f7549\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.261218 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk6gm\" (UniqueName: \"kubernetes.io/projected/943ee0ba-2610-4d88-b211-4ddf83cb5b03-kube-api-access-bk6gm\") pod \"keystone-operator-controller-manager-5bd55b4bff-vq46v\" (UID: \"943ee0ba-2610-4d88-b211-4ddf83cb5b03\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.261240 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txpxg\" (UniqueName: \"kubernetes.io/projected/2e07068f-3f60-4b35-9465-1adb7cda6525-kube-api-access-txpxg\") pod \"ironic-operator-controller-manager-7975b88857-p75vq\" (UID: \"2e07068f-3f60-4b35-9465-1adb7cda6525\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.261257 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert\") pod \"infra-operator-controller-manager-7d857cc749-rzjpm\" (UID: \"cdca0d27-1d59-4858-85a4-681c1d6abf90\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:50 crc kubenswrapper[4799]: E0930 14:35:50.264729 4799 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Sep 30 14:35:50 crc kubenswrapper[4799]: E0930 14:35:50.264836 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert podName:cdca0d27-1d59-4858-85a4-681c1d6abf90 nodeName:}" failed. No retries permitted until 2025-09-30 14:35:50.764811802 +0000 UTC m=+972.848412229 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert") pod "infra-operator-controller-manager-7d857cc749-rzjpm" (UID: "cdca0d27-1d59-4858-85a4-681c1d6abf90") : secret "infra-operator-webhook-server-cert" not found Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.299009 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.300181 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.307988 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.312120 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9ql8d" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.316036 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.317407 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.321557 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-zsfn4" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.362489 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-858s6\" (UniqueName: \"kubernetes.io/projected/4ecfaaa8-a63c-4f41-a08b-5d497a71e16d-kube-api-access-858s6\") pod \"nova-operator-controller-manager-c7c776c96-vwjtl\" (UID: \"4ecfaaa8-a63c-4f41-a08b-5d497a71e16d\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.362550 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk9qc\" (UniqueName: \"kubernetes.io/projected/c8d8a2c1-6828-49d3-bd9a-baa3fe58790e-kube-api-access-xk9qc\") pod \"mariadb-operator-controller-manager-88c7-pxd4j\" (UID: \"c8d8a2c1-6828-49d3-bd9a-baa3fe58790e\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.362582 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgnn7\" (UniqueName: \"kubernetes.io/projected/40600429-6eda-47b1-bf61-0e22189f7549-kube-api-access-pgnn7\") pod \"manila-operator-controller-manager-6d68dbc695-2jnc8\" (UID: \"40600429-6eda-47b1-bf61-0e22189f7549\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.362613 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk6gm\" (UniqueName: \"kubernetes.io/projected/943ee0ba-2610-4d88-b211-4ddf83cb5b03-kube-api-access-bk6gm\") pod \"keystone-operator-controller-manager-5bd55b4bff-vq46v\" (UID: \"943ee0ba-2610-4d88-b211-4ddf83cb5b03\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.362703 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4q8m\" (UniqueName: \"kubernetes.io/projected/f3e1a985-b655-4dc8-8e5f-f49a6459f512-kube-api-access-d4q8m\") pod \"neutron-operator-controller-manager-64d7b59854-9mvns\" (UID: \"f3e1a985-b655-4dc8-8e5f-f49a6459f512\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.363825 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.371034 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.372674 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5xzr\" (UniqueName: \"kubernetes.io/projected/cdca0d27-1d59-4858-85a4-681c1d6abf90-kube-api-access-p5xzr\") pod \"infra-operator-controller-manager-7d857cc749-rzjpm\" (UID: \"cdca0d27-1d59-4858-85a4-681c1d6abf90\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.375436 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv2cx\" (UniqueName: \"kubernetes.io/projected/fdb1977f-1341-46ae-a4db-e78e20b6d5bd-kube-api-access-rv2cx\") pod \"horizon-operator-controller-manager-9f4696d94-k955x\" (UID: \"fdb1977f-1341-46ae-a4db-e78e20b6d5bd\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.375843 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txpxg\" (UniqueName: \"kubernetes.io/projected/2e07068f-3f60-4b35-9465-1adb7cda6525-kube-api-access-txpxg\") pod \"ironic-operator-controller-manager-7975b88857-p75vq\" (UID: \"2e07068f-3f60-4b35-9465-1adb7cda6525\") " pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.409160 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.410345 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.420574 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk6gm\" (UniqueName: \"kubernetes.io/projected/943ee0ba-2610-4d88-b211-4ddf83cb5b03-kube-api-access-bk6gm\") pod \"keystone-operator-controller-manager-5bd55b4bff-vq46v\" (UID: \"943ee0ba-2610-4d88-b211-4ddf83cb5b03\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.429112 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.433874 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.434414 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.442993 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.460245 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.460390 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-tsl6n" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.472324 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgnn7\" (UniqueName: \"kubernetes.io/projected/40600429-6eda-47b1-bf61-0e22189f7549-kube-api-access-pgnn7\") pod \"manila-operator-controller-manager-6d68dbc695-2jnc8\" (UID: \"40600429-6eda-47b1-bf61-0e22189f7549\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.472867 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.479135 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-xl24h" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.480208 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-858s6\" (UniqueName: \"kubernetes.io/projected/4ecfaaa8-a63c-4f41-a08b-5d497a71e16d-kube-api-access-858s6\") pod \"nova-operator-controller-manager-c7c776c96-vwjtl\" (UID: \"4ecfaaa8-a63c-4f41-a08b-5d497a71e16d\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.481308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4q8m\" (UniqueName: \"kubernetes.io/projected/f3e1a985-b655-4dc8-8e5f-f49a6459f512-kube-api-access-d4q8m\") pod \"neutron-operator-controller-manager-64d7b59854-9mvns\" (UID: \"f3e1a985-b655-4dc8-8e5f-f49a6459f512\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.481358 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9nsp\" (UniqueName: \"kubernetes.io/projected/34ffd2cc-7377-4160-b03d-cae022be0262-kube-api-access-r9nsp\") pod \"octavia-operator-controller-manager-76fcc6dc7c-rpsrk\" (UID: \"34ffd2cc-7377-4160-b03d-cae022be0262\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.482167 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk9qc\" (UniqueName: \"kubernetes.io/projected/c8d8a2c1-6828-49d3-bd9a-baa3fe58790e-kube-api-access-xk9qc\") pod \"mariadb-operator-controller-manager-88c7-pxd4j\" (UID: \"c8d8a2c1-6828-49d3-bd9a-baa3fe58790e\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.499148 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.500683 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.527817 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.529391 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4q8m\" (UniqueName: \"kubernetes.io/projected/f3e1a985-b655-4dc8-8e5f-f49a6459f512-kube-api-access-d4q8m\") pod \"neutron-operator-controller-manager-64d7b59854-9mvns\" (UID: \"f3e1a985-b655-4dc8-8e5f-f49a6459f512\") " pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.538349 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-858s6\" (UniqueName: \"kubernetes.io/projected/4ecfaaa8-a63c-4f41-a08b-5d497a71e16d-kube-api-access-858s6\") pod \"nova-operator-controller-manager-c7c776c96-vwjtl\" (UID: \"4ecfaaa8-a63c-4f41-a08b-5d497a71e16d\") " pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.550342 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-bc9zh" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.580176 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.585119 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-q7tzd\" (UID: \"68487038-29af-45ea-8619-acd75f3d9a7d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.585181 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmxsb\" (UniqueName: \"kubernetes.io/projected/ee68b34b-53e7-452e-a113-ed35ba5c0b85-kube-api-access-vmxsb\") pod \"ovn-operator-controller-manager-9976ff44c-qrwt4\" (UID: \"ee68b34b-53e7-452e-a113-ed35ba5c0b85\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.585280 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7z6\" (UniqueName: \"kubernetes.io/projected/68487038-29af-45ea-8619-acd75f3d9a7d-kube-api-access-wf7z6\") pod \"openstack-baremetal-operator-controller-manager-6d776955-q7tzd\" (UID: \"68487038-29af-45ea-8619-acd75f3d9a7d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.585543 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9nsp\" (UniqueName: \"kubernetes.io/projected/34ffd2cc-7377-4160-b03d-cae022be0262-kube-api-access-r9nsp\") pod \"octavia-operator-controller-manager-76fcc6dc7c-rpsrk\" (UID: \"34ffd2cc-7377-4160-b03d-cae022be0262\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.588312 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.598055 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.620243 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.640257 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.674568 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9nsp\" (UniqueName: \"kubernetes.io/projected/34ffd2cc-7377-4160-b03d-cae022be0262-kube-api-access-r9nsp\") pod \"octavia-operator-controller-manager-76fcc6dc7c-rpsrk\" (UID: \"34ffd2cc-7377-4160-b03d-cae022be0262\") " pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.684356 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.687032 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-q7tzd\" (UID: \"68487038-29af-45ea-8619-acd75f3d9a7d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.687322 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmxsb\" (UniqueName: \"kubernetes.io/projected/ee68b34b-53e7-452e-a113-ed35ba5c0b85-kube-api-access-vmxsb\") pod \"ovn-operator-controller-manager-9976ff44c-qrwt4\" (UID: \"ee68b34b-53e7-452e-a113-ed35ba5c0b85\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.687476 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7z6\" (UniqueName: \"kubernetes.io/projected/68487038-29af-45ea-8619-acd75f3d9a7d-kube-api-access-wf7z6\") pod \"openstack-baremetal-operator-controller-manager-6d776955-q7tzd\" (UID: \"68487038-29af-45ea-8619-acd75f3d9a7d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:50 crc kubenswrapper[4799]: E0930 14:35:50.688630 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 14:35:50 crc kubenswrapper[4799]: E0930 14:35:50.688791 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert podName:68487038-29af-45ea-8619-acd75f3d9a7d nodeName:}" failed. No retries permitted until 2025-09-30 14:35:51.188775446 +0000 UTC m=+973.272375873 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-q7tzd" (UID: "68487038-29af-45ea-8619-acd75f3d9a7d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.705744 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.722370 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.743420 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.743916 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmxsb\" (UniqueName: \"kubernetes.io/projected/ee68b34b-53e7-452e-a113-ed35ba5c0b85-kube-api-access-vmxsb\") pod \"ovn-operator-controller-manager-9976ff44c-qrwt4\" (UID: \"ee68b34b-53e7-452e-a113-ed35ba5c0b85\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.755516 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-zcfnw" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.780824 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.795072 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.796333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8w5j\" (UniqueName: \"kubernetes.io/projected/b9a9cd33-4400-4568-822f-effb32929bb3-kube-api-access-h8w5j\") pod \"placement-operator-controller-manager-589c58c6c-sz2rp\" (UID: \"b9a9cd33-4400-4568-822f-effb32929bb3\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.796420 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert\") pod \"infra-operator-controller-manager-7d857cc749-rzjpm\" (UID: \"cdca0d27-1d59-4858-85a4-681c1d6abf90\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:50 crc kubenswrapper[4799]: E0930 14:35:50.796606 4799 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Sep 30 14:35:50 crc kubenswrapper[4799]: E0930 14:35:50.796676 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert podName:cdca0d27-1d59-4858-85a4-681c1d6abf90 nodeName:}" failed. No retries permitted until 2025-09-30 14:35:51.796643709 +0000 UTC m=+973.880244136 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert") pod "infra-operator-controller-manager-7d857cc749-rzjpm" (UID: "cdca0d27-1d59-4858-85a4-681c1d6abf90") : secret "infra-operator-webhook-server-cert" not found Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.807545 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7z6\" (UniqueName: \"kubernetes.io/projected/68487038-29af-45ea-8619-acd75f3d9a7d-kube-api-access-wf7z6\") pod \"openstack-baremetal-operator-controller-manager-6d776955-q7tzd\" (UID: \"68487038-29af-45ea-8619-acd75f3d9a7d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.818383 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.825440 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.827464 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.833202 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wfzhk" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.844440 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.846127 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.851558 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-9bpkq" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.860627 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.867031 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.886974 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.906195 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8w5j\" (UniqueName: \"kubernetes.io/projected/b9a9cd33-4400-4568-822f-effb32929bb3-kube-api-access-h8w5j\") pod \"placement-operator-controller-manager-589c58c6c-sz2rp\" (UID: \"b9a9cd33-4400-4568-822f-effb32929bb3\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.906313 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpgrk\" (UniqueName: \"kubernetes.io/projected/208f263c-4523-45e3-bf7e-19c8407bc599-kube-api-access-xpgrk\") pod \"swift-operator-controller-manager-bc7dc7bd9-68qzx\" (UID: \"208f263c-4523-45e3-bf7e-19c8407bc599\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.906362 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgqwh\" (UniqueName: \"kubernetes.io/projected/3a3cf54d-3dce-46a0-873c-412159358f6a-kube-api-access-pgqwh\") pod \"telemetry-operator-controller-manager-b8d54b5d7-9mwxx\" (UID: \"3a3cf54d-3dce-46a0-873c-412159358f6a\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.952850 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.954218 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.966355 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-lx8wr" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.983104 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9"] Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.989930 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.993214 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8w5j\" (UniqueName: \"kubernetes.io/projected/b9a9cd33-4400-4568-822f-effb32929bb3-kube-api-access-h8w5j\") pod \"placement-operator-controller-manager-589c58c6c-sz2rp\" (UID: \"b9a9cd33-4400-4568-822f-effb32929bb3\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" Sep 30 14:35:50 crc kubenswrapper[4799]: I0930 14:35:50.996051 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bfzjd" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.016934 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8qln\" (UniqueName: \"kubernetes.io/projected/26c4c0df-9057-493d-9a60-b39c664991a4-kube-api-access-n8qln\") pod \"watcher-operator-controller-manager-76669f99c-c7mtd\" (UID: \"26c4c0df-9057-493d-9a60-b39c664991a4\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.017089 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpgrk\" (UniqueName: \"kubernetes.io/projected/208f263c-4523-45e3-bf7e-19c8407bc599-kube-api-access-xpgrk\") pod \"swift-operator-controller-manager-bc7dc7bd9-68qzx\" (UID: \"208f263c-4523-45e3-bf7e-19c8407bc599\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.017183 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgqwh\" (UniqueName: \"kubernetes.io/projected/3a3cf54d-3dce-46a0-873c-412159358f6a-kube-api-access-pgqwh\") pod \"telemetry-operator-controller-manager-b8d54b5d7-9mwxx\" (UID: \"3a3cf54d-3dce-46a0-873c-412159358f6a\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.017896 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd"] Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.072908 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpgrk\" (UniqueName: \"kubernetes.io/projected/208f263c-4523-45e3-bf7e-19c8407bc599-kube-api-access-xpgrk\") pod \"swift-operator-controller-manager-bc7dc7bd9-68qzx\" (UID: \"208f263c-4523-45e3-bf7e-19c8407bc599\") " pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.075701 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9"] Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.132379 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8qln\" (UniqueName: \"kubernetes.io/projected/26c4c0df-9057-493d-9a60-b39c664991a4-kube-api-access-n8qln\") pod \"watcher-operator-controller-manager-76669f99c-c7mtd\" (UID: \"26c4c0df-9057-493d-9a60-b39c664991a4\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.132854 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh94l\" (UniqueName: \"kubernetes.io/projected/bb413bd9-a2c6-43c9-95d4-c59bbc08e930-kube-api-access-wh94l\") pod \"test-operator-controller-manager-f66b554c6-nxbv9\" (UID: \"bb413bd9-a2c6-43c9-95d4-c59bbc08e930\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.167288 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgqwh\" (UniqueName: \"kubernetes.io/projected/3a3cf54d-3dce-46a0-873c-412159358f6a-kube-api-access-pgqwh\") pod \"telemetry-operator-controller-manager-b8d54b5d7-9mwxx\" (UID: \"3a3cf54d-3dce-46a0-873c-412159358f6a\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.169797 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8qln\" (UniqueName: \"kubernetes.io/projected/26c4c0df-9057-493d-9a60-b39c664991a4-kube-api-access-n8qln\") pod \"watcher-operator-controller-manager-76669f99c-c7mtd\" (UID: \"26c4c0df-9057-493d-9a60-b39c664991a4\") " pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.212538 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.247622 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-q7tzd\" (UID: \"68487038-29af-45ea-8619-acd75f3d9a7d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.248181 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.250923 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh94l\" (UniqueName: \"kubernetes.io/projected/bb413bd9-a2c6-43c9-95d4-c59bbc08e930-kube-api-access-wh94l\") pod \"test-operator-controller-manager-f66b554c6-nxbv9\" (UID: \"bb413bd9-a2c6-43c9-95d4-c59bbc08e930\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" Sep 30 14:35:51 crc kubenswrapper[4799]: E0930 14:35:51.248409 4799 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 14:35:51 crc kubenswrapper[4799]: E0930 14:35:51.251670 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert podName:68487038-29af-45ea-8619-acd75f3d9a7d nodeName:}" failed. No retries permitted until 2025-09-30 14:35:52.25161712 +0000 UTC m=+974.335217547 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert") pod "openstack-baremetal-operator-controller-manager-6d776955-q7tzd" (UID: "68487038-29af-45ea-8619-acd75f3d9a7d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.310629 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.328494 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh94l\" (UniqueName: \"kubernetes.io/projected/bb413bd9-a2c6-43c9-95d4-c59bbc08e930-kube-api-access-wh94l\") pod \"test-operator-controller-manager-f66b554c6-nxbv9\" (UID: \"bb413bd9-a2c6-43c9-95d4-c59bbc08e930\") " pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.329123 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.335490 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh"] Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.346142 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.358355 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-mdrpn" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.358945 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.398337 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx"] Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.399419 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.403346 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-hqgng" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.408039 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.412900 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh"] Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.458507 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ece3c4-17d4-4b6f-8156-f90cfb794581-cert\") pod \"openstack-operator-controller-manager-698bb85c6f-m82xh\" (UID: \"28ece3c4-17d4-4b6f-8156-f90cfb794581\") " pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.458618 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v89nj\" (UniqueName: \"kubernetes.io/projected/28ece3c4-17d4-4b6f-8156-f90cfb794581-kube-api-access-v89nj\") pod \"openstack-operator-controller-manager-698bb85c6f-m82xh\" (UID: \"28ece3c4-17d4-4b6f-8156-f90cfb794581\") " pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.458712 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gldc2\" (UniqueName: \"kubernetes.io/projected/78d18534-fd0c-454f-80f3-b65ddd5a24c9-kube-api-access-gldc2\") pod \"rabbitmq-cluster-operator-manager-79d8469568-7tlqx\" (UID: \"78d18534-fd0c-454f-80f3-b65ddd5a24c9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.486719 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx"] Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.492548 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn"] Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.564252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gldc2\" (UniqueName: \"kubernetes.io/projected/78d18534-fd0c-454f-80f3-b65ddd5a24c9-kube-api-access-gldc2\") pod \"rabbitmq-cluster-operator-manager-79d8469568-7tlqx\" (UID: \"78d18534-fd0c-454f-80f3-b65ddd5a24c9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.564418 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ece3c4-17d4-4b6f-8156-f90cfb794581-cert\") pod \"openstack-operator-controller-manager-698bb85c6f-m82xh\" (UID: \"28ece3c4-17d4-4b6f-8156-f90cfb794581\") " pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:51 crc kubenswrapper[4799]: E0930 14:35:51.566672 4799 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Sep 30 14:35:51 crc kubenswrapper[4799]: E0930 14:35:51.566807 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28ece3c4-17d4-4b6f-8156-f90cfb794581-cert podName:28ece3c4-17d4-4b6f-8156-f90cfb794581 nodeName:}" failed. No retries permitted until 2025-09-30 14:35:52.066790074 +0000 UTC m=+974.150390511 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/28ece3c4-17d4-4b6f-8156-f90cfb794581-cert") pod "openstack-operator-controller-manager-698bb85c6f-m82xh" (UID: "28ece3c4-17d4-4b6f-8156-f90cfb794581") : secret "webhook-server-cert" not found Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.593727 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v89nj\" (UniqueName: \"kubernetes.io/projected/28ece3c4-17d4-4b6f-8156-f90cfb794581-kube-api-access-v89nj\") pod \"openstack-operator-controller-manager-698bb85c6f-m82xh\" (UID: \"28ece3c4-17d4-4b6f-8156-f90cfb794581\") " pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:51 crc kubenswrapper[4799]: W0930 14:35:51.622082 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bc24269_6185_41a5_861b_19c88576c223.slice/crio-8958244c0912551bb76f3a951bcb42b163e800861b14ed70481726a112275070 WatchSource:0}: Error finding container 8958244c0912551bb76f3a951bcb42b163e800861b14ed70481726a112275070: Status 404 returned error can't find the container with id 8958244c0912551bb76f3a951bcb42b163e800861b14ed70481726a112275070 Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.622321 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v89nj\" (UniqueName: \"kubernetes.io/projected/28ece3c4-17d4-4b6f-8156-f90cfb794581-kube-api-access-v89nj\") pod \"openstack-operator-controller-manager-698bb85c6f-m82xh\" (UID: \"28ece3c4-17d4-4b6f-8156-f90cfb794581\") " pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.644150 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gldc2\" (UniqueName: \"kubernetes.io/projected/78d18534-fd0c-454f-80f3-b65ddd5a24c9-kube-api-access-gldc2\") pod \"rabbitmq-cluster-operator-manager-79d8469568-7tlqx\" (UID: \"78d18534-fd0c-454f-80f3-b65ddd5a24c9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.799569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert\") pod \"infra-operator-controller-manager-7d857cc749-rzjpm\" (UID: \"cdca0d27-1d59-4858-85a4-681c1d6abf90\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.804339 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.814425 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cdca0d27-1d59-4858-85a4-681c1d6abf90-cert\") pod \"infra-operator-controller-manager-7d857cc749-rzjpm\" (UID: \"cdca0d27-1d59-4858-85a4-681c1d6abf90\") " pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.860592 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp"] Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.879406 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:35:51 crc kubenswrapper[4799]: I0930 14:35:51.966159 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq"] Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.023754 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2"] Sep 30 14:35:52 crc kubenswrapper[4799]: W0930 14:35:52.049261 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e07068f_3f60_4b35_9465_1adb7cda6525.slice/crio-4bca1a8b57d4ed25c4f0d92b1dcc1a0bc5af854bdac3a5d385548c813c1f8d7d WatchSource:0}: Error finding container 4bca1a8b57d4ed25c4f0d92b1dcc1a0bc5af854bdac3a5d385548c813c1f8d7d: Status 404 returned error can't find the container with id 4bca1a8b57d4ed25c4f0d92b1dcc1a0bc5af854bdac3a5d385548c813c1f8d7d Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.146184 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ece3c4-17d4-4b6f-8156-f90cfb794581-cert\") pod \"openstack-operator-controller-manager-698bb85c6f-m82xh\" (UID: \"28ece3c4-17d4-4b6f-8156-f90cfb794581\") " pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.157769 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ece3c4-17d4-4b6f-8156-f90cfb794581-cert\") pod \"openstack-operator-controller-manager-698bb85c6f-m82xh\" (UID: \"28ece3c4-17d4-4b6f-8156-f90cfb794581\") " pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.192650 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf"] Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.293487 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" event={"ID":"08704fb7-1721-4f10-97a1-f2359d01cdaf","Type":"ContainerStarted","Data":"8069cfed615dece5e75fab1e7eb5e51aa0ebe20eeac4d7f0d7ea1f2222dfd6a2"} Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.293629 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.298209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" event={"ID":"2e07068f-3f60-4b35-9465-1adb7cda6525","Type":"ContainerStarted","Data":"4bca1a8b57d4ed25c4f0d92b1dcc1a0bc5af854bdac3a5d385548c813c1f8d7d"} Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.299248 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" event={"ID":"19539cfa-6f83-4636-8193-d2373038f353","Type":"ContainerStarted","Data":"601bf847425a885d4fbcba8c397a188ced0506bb5e06f93cd8f65bee2a019062"} Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.300128 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" event={"ID":"8bc24269-6185-41a5-861b-19c88576c223","Type":"ContainerStarted","Data":"8958244c0912551bb76f3a951bcb42b163e800861b14ed70481726a112275070"} Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.301187 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" event={"ID":"b8a734fd-7d92-4b2d-9bb4-91eee56047c5","Type":"ContainerStarted","Data":"6fec14a3feab04eaaff5d6b910633fd09113408f1ee2d73e8584243b6a09a72f"} Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.350972 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-q7tzd\" (UID: \"68487038-29af-45ea-8619-acd75f3d9a7d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.366818 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68487038-29af-45ea-8619-acd75f3d9a7d-cert\") pod \"openstack-baremetal-operator-controller-manager-6d776955-q7tzd\" (UID: \"68487038-29af-45ea-8619-acd75f3d9a7d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.402312 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.441159 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v"] Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.449751 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j"] Sep 30 14:35:52 crc kubenswrapper[4799]: W0930 14:35:52.470518 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod943ee0ba_2610_4d88_b211_4ddf83cb5b03.slice/crio-b5a013b3f893d508ddb34190a7d1e838ab070c4bc7a9638347ba9b7fae7c367a WatchSource:0}: Error finding container b5a013b3f893d508ddb34190a7d1e838ab070c4bc7a9638347ba9b7fae7c367a: Status 404 returned error can't find the container with id b5a013b3f893d508ddb34190a7d1e838ab070c4bc7a9638347ba9b7fae7c367a Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.493307 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns"] Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.580048 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp"] Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.597696 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl"] Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.663051 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8"] Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.746594 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4"] Sep 30 14:35:52 crc kubenswrapper[4799]: I0930 14:35:52.778159 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk"] Sep 30 14:35:52 crc kubenswrapper[4799]: W0930 14:35:52.837045 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee68b34b_53e7_452e_a113_ed35ba5c0b85.slice/crio-5d802dac8f4dbc3162edb8994439b61b115f6c6a73345be545a5c09bcd809a35 WatchSource:0}: Error finding container 5d802dac8f4dbc3162edb8994439b61b115f6c6a73345be545a5c09bcd809a35: Status 404 returned error can't find the container with id 5d802dac8f4dbc3162edb8994439b61b115f6c6a73345be545a5c09bcd809a35 Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.163024 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx"] Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.209060 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm"] Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.215099 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx"] Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.234112 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx"] Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.262514 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9"] Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.320093 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd"] Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.354367 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x"] Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.375295 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp"] Sep 30 14:35:53 crc kubenswrapper[4799]: E0930 14:35:53.378917 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wh94l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-f66b554c6-nxbv9_openstack-operators(bb413bd9-a2c6-43c9-95d4-c59bbc08e930): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:35:53 crc kubenswrapper[4799]: E0930 14:35:53.379162 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n8qln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-76669f99c-c7mtd_openstack-operators(26c4c0df-9057-493d-9a60-b39c664991a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.391386 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh"] Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.394504 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" event={"ID":"4ecfaaa8-a63c-4f41-a08b-5d497a71e16d","Type":"ContainerStarted","Data":"dbc7db9469c842c8d17308488ab5677d04b7afba827add03fbfd26fd7b0c0bab"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.410525 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" event={"ID":"e0002170-fb28-4c87-a970-350f92c891bd","Type":"ContainerStarted","Data":"918f40b266cd7d2b85bd48c7968867792ac26eaadb737b2943390f414a09b577"} Sep 30 14:35:53 crc kubenswrapper[4799]: W0930 14:35:53.413730 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9a9cd33_4400_4568_822f_effb32929bb3.slice/crio-a9f08426c163a3696feecfbe04af7a7ae5fb1137bb71332cbded390940773aab WatchSource:0}: Error finding container a9f08426c163a3696feecfbe04af7a7ae5fb1137bb71332cbded390940773aab: Status 404 returned error can't find the container with id a9f08426c163a3696feecfbe04af7a7ae5fb1137bb71332cbded390940773aab Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.419988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" event={"ID":"943ee0ba-2610-4d88-b211-4ddf83cb5b03","Type":"ContainerStarted","Data":"b5a013b3f893d508ddb34190a7d1e838ab070c4bc7a9638347ba9b7fae7c367a"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.423796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" event={"ID":"40600429-6eda-47b1-bf61-0e22189f7549","Type":"ContainerStarted","Data":"581a7ef5ea36d7235cd14b62cb06352b905ff4a267ddb61fd92c1511a99cadea"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.428504 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" event={"ID":"ee68b34b-53e7-452e-a113-ed35ba5c0b85","Type":"ContainerStarted","Data":"5d802dac8f4dbc3162edb8994439b61b115f6c6a73345be545a5c09bcd809a35"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.430253 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" event={"ID":"3a3cf54d-3dce-46a0-873c-412159358f6a","Type":"ContainerStarted","Data":"bffd38fd85a556a680440a4a49920c2f46b5fdedbfb272cb7a47c446434959be"} Sep 30 14:35:53 crc kubenswrapper[4799]: E0930 14:35:53.432967 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h8w5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-sz2rp_openstack-operators(b9a9cd33-4400-4568-822f-effb32929bb3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.436468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" event={"ID":"bb413bd9-a2c6-43c9-95d4-c59bbc08e930","Type":"ContainerStarted","Data":"fda4057c0c9f05cf5fdf83c798a7b48c1cebdfea5d2340f82af4b0a530393551"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.442462 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" event={"ID":"78d18534-fd0c-454f-80f3-b65ddd5a24c9","Type":"ContainerStarted","Data":"087fc75c1accd80ec3bc176173c3f8f26fe9f901816809472919d12f6c305523"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.444629 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" event={"ID":"c8d8a2c1-6828-49d3-bd9a-baa3fe58790e","Type":"ContainerStarted","Data":"bf71ef4030cd1433ee9a05b6e76dbf29e858359d54de177918015507d625612f"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.446988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" event={"ID":"34ffd2cc-7377-4160-b03d-cae022be0262","Type":"ContainerStarted","Data":"13520efbcc9ecc48eaff485d84f1b00d506df0fa25a8b2f32ceff0a4a7433483"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.451186 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" event={"ID":"208f263c-4523-45e3-bf7e-19c8407bc599","Type":"ContainerStarted","Data":"475c26044f357802722aafefbbc74b8afdccdb10c079d44ff6b3c06c12174954"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.457497 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" event={"ID":"f3e1a985-b655-4dc8-8e5f-f49a6459f512","Type":"ContainerStarted","Data":"76a401b06a18aefc7ea28d9bd12251d7bdf377140b2b560dd44f2b133ceb10f9"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.464791 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" event={"ID":"cdca0d27-1d59-4858-85a4-681c1d6abf90","Type":"ContainerStarted","Data":"b454d969fc38df303d4d9174afe475180eb8242e0eebb81c05a66a4b3abd7d11"} Sep 30 14:35:53 crc kubenswrapper[4799]: I0930 14:35:53.574820 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd"] Sep 30 14:35:53 crc kubenswrapper[4799]: E0930 14:35:53.627275 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" podUID="bb413bd9-a2c6-43c9-95d4-c59bbc08e930" Sep 30 14:35:53 crc kubenswrapper[4799]: E0930 14:35:53.708715 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" podUID="26c4c0df-9057-493d-9a60-b39c664991a4" Sep 30 14:35:53 crc kubenswrapper[4799]: E0930 14:35:53.930219 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" podUID="b9a9cd33-4400-4568-822f-effb32929bb3" Sep 30 14:35:54 crc kubenswrapper[4799]: E0930 14:35:54.512552 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" podUID="bb413bd9-a2c6-43c9-95d4-c59bbc08e930" Sep 30 14:35:54 crc kubenswrapper[4799]: E0930 14:35:54.512808 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" podUID="b9a9cd33-4400-4568-822f-effb32929bb3" Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.513470 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" event={"ID":"bb413bd9-a2c6-43c9-95d4-c59bbc08e930","Type":"ContainerStarted","Data":"5dbe0fc3b69ff4b541f67edf4ee71b0b99c6048a3f4199fb5c622f8b3c3ef53a"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.513518 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" event={"ID":"b9a9cd33-4400-4568-822f-effb32929bb3","Type":"ContainerStarted","Data":"3f930fbdd955207c61779eb2e52ef7638298a2eb663d5c03851476696129e8be"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.513538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" event={"ID":"b9a9cd33-4400-4568-822f-effb32929bb3","Type":"ContainerStarted","Data":"a9f08426c163a3696feecfbe04af7a7ae5fb1137bb71332cbded390940773aab"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.535454 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" event={"ID":"26c4c0df-9057-493d-9a60-b39c664991a4","Type":"ContainerStarted","Data":"da988cac8dd7ed8ffbc0c6f13204d926689380716a1e08b9520af55db3347d42"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.535603 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" event={"ID":"26c4c0df-9057-493d-9a60-b39c664991a4","Type":"ContainerStarted","Data":"5165423c2033b26ac038be9dab1392a1d1d72cd554f512f019de5f51a1f8738c"} Sep 30 14:35:54 crc kubenswrapper[4799]: E0930 14:35:54.544045 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" podUID="26c4c0df-9057-493d-9a60-b39c664991a4" Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.546398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" event={"ID":"68487038-29af-45ea-8619-acd75f3d9a7d","Type":"ContainerStarted","Data":"f562c528bf5241bc343f19a3b2cc3e253080fa7bd087ad232aa3d39f99b0823c"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.550111 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" event={"ID":"fdb1977f-1341-46ae-a4db-e78e20b6d5bd","Type":"ContainerStarted","Data":"34e9409ce38e87df75556c28f3f4fb02c1cf437a14d82fe00834236d7a5b35c5"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.552357 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" event={"ID":"28ece3c4-17d4-4b6f-8156-f90cfb794581","Type":"ContainerStarted","Data":"1374911ad00dd0a93e21062507d727d9216db6408e8998882b37b913c0a3b79d"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.552397 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" event={"ID":"28ece3c4-17d4-4b6f-8156-f90cfb794581","Type":"ContainerStarted","Data":"5c0c8ac3050b501349ddbcac8b12d83e9d1fed26ed99fa276247807be39e1c6e"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.552408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" event={"ID":"28ece3c4-17d4-4b6f-8156-f90cfb794581","Type":"ContainerStarted","Data":"8cf838bc55be650472dbe25dcc6043a53c639862ed61dbd41107fc497e3d19d0"} Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.552675 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:35:54 crc kubenswrapper[4799]: I0930 14:35:54.664953 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" podStartSLOduration=4.664896414 podStartE2EDuration="4.664896414s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:35:54.661059772 +0000 UTC m=+976.744660209" watchObservedRunningTime="2025-09-30 14:35:54.664896414 +0000 UTC m=+976.748496841" Sep 30 14:35:55 crc kubenswrapper[4799]: E0930 14:35:55.608145 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7169dfadf5f5589f14ca52700d2eba991c2a0c7733f6a1ea795752d993d7f61b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" podUID="26c4c0df-9057-493d-9a60-b39c664991a4" Sep 30 14:35:55 crc kubenswrapper[4799]: E0930 14:35:55.608932 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" podUID="b9a9cd33-4400-4568-822f-effb32929bb3" Sep 30 14:35:55 crc kubenswrapper[4799]: E0930 14:35:55.609067 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:a303e460aec09217f90043b8ff19c01061af003b614833b33a593df9c00ddf80\\\"\"" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" podUID="bb413bd9-a2c6-43c9-95d4-c59bbc08e930" Sep 30 14:36:02 crc kubenswrapper[4799]: I0930 14:36:02.301274 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-698bb85c6f-m82xh" Sep 30 14:36:06 crc kubenswrapper[4799]: E0930 14:36:06.916692 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c" Sep 30 14:36:06 crc kubenswrapper[4799]: E0930 14:36:06.917421 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zmqkr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5d889d78cf-8vxq2_openstack-operators(08704fb7-1721-4f10-97a1-f2359d01cdaf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:07 crc kubenswrapper[4799]: E0930 14:36:07.491523 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397" Sep 30 14:36:07 crc kubenswrapper[4799]: E0930 14:36:07.491840 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rv2cx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-9f4696d94-k955x_openstack-operators(fdb1977f-1341-46ae-a4db-e78e20b6d5bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:10 crc kubenswrapper[4799]: E0930 14:36:10.333939 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c" Sep 30 14:36:10 crc kubenswrapper[4799]: E0930 14:36:10.334275 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xpgrk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-bc7dc7bd9-68qzx_openstack-operators(208f263c-4523-45e3-bf7e-19c8407bc599): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:10 crc kubenswrapper[4799]: E0930 14:36:10.860833 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:4d08afd31dc5ded10c54a5541f514ac351e9b40a183285b3db27d0757a6354c8" Sep 30 14:36:10 crc kubenswrapper[4799]: E0930 14:36:10.862767 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:4d08afd31dc5ded10c54a5541f514ac351e9b40a183285b3db27d0757a6354c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r9nsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-76fcc6dc7c-rpsrk_openstack-operators(34ffd2cc-7377-4160-b03d-cae022be0262): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:11 crc kubenswrapper[4799]: E0930 14:36:11.402756 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1" Sep 30 14:36:11 crc kubenswrapper[4799]: E0930 14:36:11.402918 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tfdpn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-644bddb6d8-kqfcp_openstack-operators(e0002170-fb28-4c87-a970-350f92c891bd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:11 crc kubenswrapper[4799]: E0930 14:36:11.863236 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:f6b935f67979298c3c263ad84d277e5cf26c0dbba3f85f255c1ec4d1d75241d2" Sep 30 14:36:11 crc kubenswrapper[4799]: E0930 14:36:11.863427 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:f6b935f67979298c3c263ad84d277e5cf26c0dbba3f85f255c1ec4d1d75241d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jr9hs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-84f4f7b77b-mplqn_openstack-operators(8bc24269-6185-41a5-861b-19c88576c223): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:13 crc kubenswrapper[4799]: E0930 14:36:13.547237 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:485df5c7813cdf4cf21f48ec48c8e3e4962fee6a1ae4c64f7af127d5ab346a10" Sep 30 14:36:13 crc kubenswrapper[4799]: E0930 14:36:13.549001 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:485df5c7813cdf4cf21f48ec48c8e3e4962fee6a1ae4c64f7af127d5ab346a10,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d4q8m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-64d7b59854-9mvns_openstack-operators(f3e1a985-b655-4dc8-8e5f-f49a6459f512): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:14 crc kubenswrapper[4799]: E0930 14:36:14.019614 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72" Sep 30 14:36:14 crc kubenswrapper[4799]: E0930 14:36:14.019864 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r9jsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-84958c4d49-gmxsp_openstack-operators(19539cfa-6f83-4636-8193-d2373038f353): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:15 crc kubenswrapper[4799]: E0930 14:36:15.993392 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6" Sep 30 14:36:15 crc kubenswrapper[4799]: E0930 14:36:15.994317 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_LIGHTSPEED_IMAGE_URL_DEFAULT,Value:quay.io/openstack-lightspeed/rag-content:os-docs-2024.2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wf7z6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6d776955-q7tzd_openstack-operators(68487038-29af-45ea-8619-acd75f3d9a7d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:16 crc kubenswrapper[4799]: E0930 14:36:16.387785 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b" Sep 30 14:36:16 crc kubenswrapper[4799]: E0930 14:36:16.387999 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gldc2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-79d8469568-7tlqx_openstack-operators(78d18534-fd0c-454f-80f3-b65ddd5a24c9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:36:16 crc kubenswrapper[4799]: E0930 14:36:16.389824 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" podUID="78d18534-fd0c-454f-80f3-b65ddd5a24c9" Sep 30 14:36:16 crc kubenswrapper[4799]: E0930 14:36:16.783075 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" podUID="78d18534-fd0c-454f-80f3-b65ddd5a24c9" Sep 30 14:36:17 crc kubenswrapper[4799]: E0930 14:36:17.882490 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" podUID="08704fb7-1721-4f10-97a1-f2359d01cdaf" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.240682 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" podUID="fdb1977f-1341-46ae-a4db-e78e20b6d5bd" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.243207 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" podUID="e0002170-fb28-4c87-a970-350f92c891bd" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.284619 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" podUID="8bc24269-6185-41a5-861b-19c88576c223" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.337431 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" podUID="34ffd2cc-7377-4160-b03d-cae022be0262" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.360328 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" podUID="208f263c-4523-45e3-bf7e-19c8407bc599" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.384321 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" podUID="f3e1a985-b655-4dc8-8e5f-f49a6459f512" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.443687 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" podUID="19539cfa-6f83-4636-8193-d2373038f353" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.473154 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" podUID="68487038-29af-45ea-8619-acd75f3d9a7d" Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.806036 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" event={"ID":"19539cfa-6f83-4636-8193-d2373038f353","Type":"ContainerStarted","Data":"c35278bdfed05cbc39b846a03cd8aa0f6a6fab2f9d7bfacb0f64f7a479cb119a"} Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.823162 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" event={"ID":"f3e1a985-b655-4dc8-8e5f-f49a6459f512","Type":"ContainerStarted","Data":"bdcf47ad8921950654e1db5b0da89591b88d1c5f5a790c3c74802400686157f1"} Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.824538 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:485df5c7813cdf4cf21f48ec48c8e3e4962fee6a1ae4c64f7af127d5ab346a10\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" podUID="f3e1a985-b655-4dc8-8e5f-f49a6459f512" Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.826613 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" event={"ID":"b9a9cd33-4400-4568-822f-effb32929bb3","Type":"ContainerStarted","Data":"a28920e017ab5db7d297438cf3d9118afa740486e1a650d42ef7f6f716871c73"} Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.826882 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.827508 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72\\\"\"" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" podUID="19539cfa-6f83-4636-8193-d2373038f353" Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.850020 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" event={"ID":"68487038-29af-45ea-8619-acd75f3d9a7d","Type":"ContainerStarted","Data":"1955c8af28f87c9ec4505e3137286a782238204099faea39c109300fc2f92166"} Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.856044 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" podUID="68487038-29af-45ea-8619-acd75f3d9a7d" Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.867071 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" event={"ID":"4ecfaaa8-a63c-4f41-a08b-5d497a71e16d","Type":"ContainerStarted","Data":"9256cbc64cb7002e45a2b47d624fef031d7cd8e6725b93a37b5e2c1b8ca0cc24"} Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.881354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" event={"ID":"e0002170-fb28-4c87-a970-350f92c891bd","Type":"ContainerStarted","Data":"d526474ea92e2e18285430ce1351e4b5f51192af807db697c373928d5021dcf7"} Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.889961 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" podUID="e0002170-fb28-4c87-a970-350f92c891bd" Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.900907 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" event={"ID":"c8d8a2c1-6828-49d3-bd9a-baa3fe58790e","Type":"ContainerStarted","Data":"1f5e1ff2a03635c1e2290095a89c1f0a35396e15819a40ad5b04c2a887fd78f5"} Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.919507 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" event={"ID":"8bc24269-6185-41a5-861b-19c88576c223","Type":"ContainerStarted","Data":"dc1fef9686d4d1f491d73fa7c537a7a4b0b4799460c50bb20bc2020d78725c46"} Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.965817 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" event={"ID":"943ee0ba-2610-4d88-b211-4ddf83cb5b03","Type":"ContainerStarted","Data":"ce570c19a09e6e4f17891cce0d487b6c98c35b539bbff45bc43b5a3138ab1c75"} Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.981232 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:f6b935f67979298c3c263ad84d277e5cf26c0dbba3f85f255c1ec4d1d75241d2\\\"\"" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" podUID="8bc24269-6185-41a5-861b-19c88576c223" Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.984557 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" event={"ID":"ee68b34b-53e7-452e-a113-ed35ba5c0b85","Type":"ContainerStarted","Data":"ad8782de8dc88d0c73f885e70cdadb087030e8efa18d82fb22cbd951f2028a54"} Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.988449 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" podStartSLOduration=4.465937578 podStartE2EDuration="28.988428467s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.432801805 +0000 UTC m=+975.516402232" lastFinishedPulling="2025-09-30 14:36:17.955292694 +0000 UTC m=+1000.038893121" observedRunningTime="2025-09-30 14:36:18.983277896 +0000 UTC m=+1001.066878323" watchObservedRunningTime="2025-09-30 14:36:18.988428467 +0000 UTC m=+1001.072028894" Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.993403 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" event={"ID":"208f263c-4523-45e3-bf7e-19c8407bc599","Type":"ContainerStarted","Data":"19a0a4265dd8b56930fccb22dba02117fd61b7496161ecfe1527be50a1ab9fdb"} Sep 30 14:36:18 crc kubenswrapper[4799]: I0930 14:36:18.995368 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" event={"ID":"3a3cf54d-3dce-46a0-873c-412159358f6a","Type":"ContainerStarted","Data":"a21c5d814ca2abac6e98ce816aac0c5853896a84fa9ac6a751b2ef60cf4c31cd"} Sep 30 14:36:18 crc kubenswrapper[4799]: E0930 14:36:18.995563 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" podUID="208f263c-4523-45e3-bf7e-19c8407bc599" Sep 30 14:36:19 crc kubenswrapper[4799]: I0930 14:36:19.018818 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" event={"ID":"bb413bd9-a2c6-43c9-95d4-c59bbc08e930","Type":"ContainerStarted","Data":"8c58bb593a120a2aabbb1f1aa0726d51fb17e27b8b7e3846c46ea14982b5b60d"} Sep 30 14:36:19 crc kubenswrapper[4799]: I0930 14:36:19.019357 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" Sep 30 14:36:19 crc kubenswrapper[4799]: I0930 14:36:19.035035 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" event={"ID":"08704fb7-1721-4f10-97a1-f2359d01cdaf","Type":"ContainerStarted","Data":"320d08e910c063ed59b32feadb41a9f6d157c5cc719dbad8535cd2494db073b5"} Sep 30 14:36:19 crc kubenswrapper[4799]: I0930 14:36:19.045940 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" event={"ID":"34ffd2cc-7377-4160-b03d-cae022be0262","Type":"ContainerStarted","Data":"c629328bac651b5e30d4d7aa7437bcb8c8a93ebdcfdc848b844a5492c4578549"} Sep 30 14:36:19 crc kubenswrapper[4799]: E0930 14:36:19.052102 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:4d08afd31dc5ded10c54a5541f514ac351e9b40a183285b3db27d0757a6354c8\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" podUID="34ffd2cc-7377-4160-b03d-cae022be0262" Sep 30 14:36:19 crc kubenswrapper[4799]: I0930 14:36:19.085049 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" event={"ID":"fdb1977f-1341-46ae-a4db-e78e20b6d5bd","Type":"ContainerStarted","Data":"0a9293802de43c135fdd0863c312ad52572f2812ec64688a1e9e54d37f38fa83"} Sep 30 14:36:19 crc kubenswrapper[4799]: I0930 14:36:19.121234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" event={"ID":"b8a734fd-7d92-4b2d-9bb4-91eee56047c5","Type":"ContainerStarted","Data":"c9db1ddcd7323eb4d56c30e9709b0e54396662a5fd438a656bc1ceabc24f46f0"} Sep 30 14:36:19 crc kubenswrapper[4799]: I0930 14:36:19.140259 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" event={"ID":"2e07068f-3f60-4b35-9465-1adb7cda6525","Type":"ContainerStarted","Data":"9158db5f9170a3d07b844e3efa0a885f9f00b1961c2a8e692d24b38fd2e30978"} Sep 30 14:36:19 crc kubenswrapper[4799]: I0930 14:36:19.248211 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" podStartSLOduration=4.7163962 podStartE2EDuration="29.248191441s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.378695363 +0000 UTC m=+975.462295790" lastFinishedPulling="2025-09-30 14:36:17.910490604 +0000 UTC m=+999.994091031" observedRunningTime="2025-09-30 14:36:19.23583441 +0000 UTC m=+1001.319434837" watchObservedRunningTime="2025-09-30 14:36:19.248191441 +0000 UTC m=+1001.331791868" Sep 30 14:36:20 crc kubenswrapper[4799]: I0930 14:36:20.148674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" event={"ID":"cdca0d27-1d59-4858-85a4-681c1d6abf90","Type":"ContainerStarted","Data":"b518113dbe408f947668bc94801e2369e6708588ac0b12fa1e4c23ea04e4745e"} Sep 30 14:36:20 crc kubenswrapper[4799]: I0930 14:36:20.150088 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" event={"ID":"26c4c0df-9057-493d-9a60-b39c664991a4","Type":"ContainerStarted","Data":"02f5c1a15b2b08a2a1558ca877fe7ce3c8aeeb568f35f4c9210bff577606b8f6"} Sep 30 14:36:20 crc kubenswrapper[4799]: I0930 14:36:20.150285 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" Sep 30 14:36:20 crc kubenswrapper[4799]: I0930 14:36:20.156387 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" event={"ID":"40600429-6eda-47b1-bf61-0e22189f7549","Type":"ContainerStarted","Data":"c7d3e414ef954d867b19d32aa254f2b7cfb703cdca1addb3b25536dfbaccfb08"} Sep 30 14:36:20 crc kubenswrapper[4799]: E0930 14:36:20.167105 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:f6b935f67979298c3c263ad84d277e5cf26c0dbba3f85f255c1ec4d1d75241d2\\\"\"" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" podUID="8bc24269-6185-41a5-861b-19c88576c223" Sep 30 14:36:20 crc kubenswrapper[4799]: E0930 14:36:20.167113 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:485df5c7813cdf4cf21f48ec48c8e3e4962fee6a1ae4c64f7af127d5ab346a10\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" podUID="f3e1a985-b655-4dc8-8e5f-f49a6459f512" Sep 30 14:36:20 crc kubenswrapper[4799]: E0930 14:36:20.167202 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:4d08afd31dc5ded10c54a5541f514ac351e9b40a183285b3db27d0757a6354c8\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" podUID="34ffd2cc-7377-4160-b03d-cae022be0262" Sep 30 14:36:20 crc kubenswrapper[4799]: E0930 14:36:20.167118 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e3f947e9034a951620a76eaf41ceec95eefcef0eacb251b10993d6820d5e1af6\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" podUID="68487038-29af-45ea-8619-acd75f3d9a7d" Sep 30 14:36:20 crc kubenswrapper[4799]: E0930 14:36:20.167245 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:1e2c65f4331a2bb568d97fbcd02e3bca2627e133a794e1e4fd13368e86ce6bd1\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" podUID="e0002170-fb28-4c87-a970-350f92c891bd" Sep 30 14:36:20 crc kubenswrapper[4799]: E0930 14:36:20.167382 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:21792a2317c0a55e40b2a02a7d5d4682b76538ed2a2e0633199aa395e60ecc72\\\"\"" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" podUID="19539cfa-6f83-4636-8193-d2373038f353" Sep 30 14:36:20 crc kubenswrapper[4799]: E0930 14:36:20.167500 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3c6f7d737e0196ec302f44354228d783ad3b210a75703dda3b39c15c01a67e8c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" podUID="208f263c-4523-45e3-bf7e-19c8407bc599" Sep 30 14:36:20 crc kubenswrapper[4799]: I0930 14:36:20.187248 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" podStartSLOduration=5.632557484 podStartE2EDuration="30.187224833s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.378637292 +0000 UTC m=+975.462237729" lastFinishedPulling="2025-09-30 14:36:17.933304651 +0000 UTC m=+1000.016905078" observedRunningTime="2025-09-30 14:36:20.182285868 +0000 UTC m=+1002.265886315" watchObservedRunningTime="2025-09-30 14:36:20.187224833 +0000 UTC m=+1002.270825260" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.170195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" event={"ID":"cdca0d27-1d59-4858-85a4-681c1d6abf90","Type":"ContainerStarted","Data":"39b05898ffa6992961d58d31c381369a96ae019d709dc5c3c8f4205517d1062d"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.171594 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.189206 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" event={"ID":"4ecfaaa8-a63c-4f41-a08b-5d497a71e16d","Type":"ContainerStarted","Data":"9f0d0fed0b63f647991e8e4dc6f2fd919c96cdb26ccebea454e83dd8aa9fb875"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.189993 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.204797 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" podStartSLOduration=7.663946483 podStartE2EDuration="32.204775869s" podCreationTimestamp="2025-09-30 14:35:49 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.272286842 +0000 UTC m=+975.355887269" lastFinishedPulling="2025-09-30 14:36:17.813116238 +0000 UTC m=+999.896716655" observedRunningTime="2025-09-30 14:36:21.198282009 +0000 UTC m=+1003.281882456" watchObservedRunningTime="2025-09-30 14:36:21.204775869 +0000 UTC m=+1003.288376306" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.219337 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" event={"ID":"2e07068f-3f60-4b35-9465-1adb7cda6525","Type":"ContainerStarted","Data":"e86afcef1afcf392bb8585e38373d101a88661123b1ac2265377ca1132f5f0f8"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.221294 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.231088 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" podStartSLOduration=7.010392991 podStartE2EDuration="31.231063267s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.613471192 +0000 UTC m=+974.697071629" lastFinishedPulling="2025-09-30 14:36:16.834141478 +0000 UTC m=+998.917741905" observedRunningTime="2025-09-30 14:36:21.226879265 +0000 UTC m=+1003.310479702" watchObservedRunningTime="2025-09-30 14:36:21.231063267 +0000 UTC m=+1003.314663694" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.243966 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" event={"ID":"c8d8a2c1-6828-49d3-bd9a-baa3fe58790e","Type":"ContainerStarted","Data":"9342f9611f334a0c577293ef4d04ac052fe80dbc5f20c0ab288d8345a3676e2b"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.244546 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.273519 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" event={"ID":"943ee0ba-2610-4d88-b211-4ddf83cb5b03","Type":"ContainerStarted","Data":"7db89e63424d3d86ea5991d4f197dc25781fe77c31db64b44ad2607859d9f8e0"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.274744 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.293521 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" podStartSLOduration=7.512455225 podStartE2EDuration="32.293498653s" podCreationTimestamp="2025-09-30 14:35:49 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.053886573 +0000 UTC m=+974.137487000" lastFinishedPulling="2025-09-30 14:36:16.834930001 +0000 UTC m=+998.918530428" observedRunningTime="2025-09-30 14:36:21.261212499 +0000 UTC m=+1003.344812946" watchObservedRunningTime="2025-09-30 14:36:21.293498653 +0000 UTC m=+1003.377099080" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.295036 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" podStartSLOduration=6.967993752 podStartE2EDuration="31.295028767s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.507992269 +0000 UTC m=+974.591592696" lastFinishedPulling="2025-09-30 14:36:16.835027274 +0000 UTC m=+998.918627711" observedRunningTime="2025-09-30 14:36:21.28997431 +0000 UTC m=+1003.373574727" watchObservedRunningTime="2025-09-30 14:36:21.295028767 +0000 UTC m=+1003.378629194" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.299902 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" event={"ID":"40600429-6eda-47b1-bf61-0e22189f7549","Type":"ContainerStarted","Data":"b7db71987dcfe4c88017e2795aa450a167a60e481529c8feaabbe0e70db37022"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.300624 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.323071 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" event={"ID":"3a3cf54d-3dce-46a0-873c-412159358f6a","Type":"ContainerStarted","Data":"268bb937bd5b7a05c768214a0e7181b1739df2aa720c3a5419d505aec3c0070d"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.324206 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.332695 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" podStartSLOduration=7.012468442 podStartE2EDuration="31.332632437s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.514476008 +0000 UTC m=+974.598076435" lastFinishedPulling="2025-09-30 14:36:16.834640003 +0000 UTC m=+998.918240430" observedRunningTime="2025-09-30 14:36:21.327899878 +0000 UTC m=+1003.411500325" watchObservedRunningTime="2025-09-30 14:36:21.332632437 +0000 UTC m=+1003.416232874" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.357008 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" event={"ID":"b8a734fd-7d92-4b2d-9bb4-91eee56047c5","Type":"ContainerStarted","Data":"da1fc538797bc55daa5000c10c6004a5ce88e24b06b7fcaea2e6899646f073b3"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.358080 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.361233 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" podStartSLOduration=7.734623784 podStartE2EDuration="31.361219612s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.208332223 +0000 UTC m=+975.291932650" lastFinishedPulling="2025-09-30 14:36:16.834928051 +0000 UTC m=+998.918528478" observedRunningTime="2025-09-30 14:36:21.357890025 +0000 UTC m=+1003.441490452" watchObservedRunningTime="2025-09-30 14:36:21.361219612 +0000 UTC m=+1003.444820039" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.362914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" event={"ID":"ee68b34b-53e7-452e-a113-ed35ba5c0b85","Type":"ContainerStarted","Data":"09a115a53ebbeaf8812c4c585ab26b322a0140394b9ac16243a63cee69bf7a40"} Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.362980 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.403040 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" podStartSLOduration=7.275994916 podStartE2EDuration="31.403019394s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.707980975 +0000 UTC m=+974.791581402" lastFinishedPulling="2025-09-30 14:36:16.835005453 +0000 UTC m=+998.918605880" observedRunningTime="2025-09-30 14:36:21.396228896 +0000 UTC m=+1003.479829343" watchObservedRunningTime="2025-09-30 14:36:21.403019394 +0000 UTC m=+1003.486619821" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.433412 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" podStartSLOduration=7.438098595 podStartE2EDuration="31.433394052s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.839667525 +0000 UTC m=+974.923267952" lastFinishedPulling="2025-09-30 14:36:16.834962982 +0000 UTC m=+998.918563409" observedRunningTime="2025-09-30 14:36:21.432502496 +0000 UTC m=+1003.516102933" watchObservedRunningTime="2025-09-30 14:36:21.433394052 +0000 UTC m=+1003.516994479" Sep 30 14:36:21 crc kubenswrapper[4799]: I0930 14:36:21.484913 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" podStartSLOduration=7.880302738 podStartE2EDuration="32.484893608s" podCreationTimestamp="2025-09-30 14:35:49 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.230464175 +0000 UTC m=+974.314064602" lastFinishedPulling="2025-09-30 14:36:16.835055045 +0000 UTC m=+998.918655472" observedRunningTime="2025-09-30 14:36:21.48324532 +0000 UTC m=+1003.566845747" watchObservedRunningTime="2025-09-30 14:36:21.484893608 +0000 UTC m=+1003.568494035" Sep 30 14:36:22 crc kubenswrapper[4799]: I0930 14:36:22.370735 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" event={"ID":"08704fb7-1721-4f10-97a1-f2359d01cdaf","Type":"ContainerStarted","Data":"ee9a4dfc39628199837cf850cfa8434fcfd5691f794af87c19d42e8b6ea2d84d"} Sep 30 14:36:22 crc kubenswrapper[4799]: I0930 14:36:22.372045 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" Sep 30 14:36:22 crc kubenswrapper[4799]: I0930 14:36:22.374597 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" event={"ID":"fdb1977f-1341-46ae-a4db-e78e20b6d5bd","Type":"ContainerStarted","Data":"8190d8f6d2d07399d653eaf991b5cc3b119f73bf50f3dd2fc99fb3808c666ba1"} Sep 30 14:36:22 crc kubenswrapper[4799]: I0930 14:36:22.412499 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" podStartSLOduration=4.343795613 podStartE2EDuration="33.412479365s" podCreationTimestamp="2025-09-30 14:35:49 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.077818413 +0000 UTC m=+974.161418830" lastFinishedPulling="2025-09-30 14:36:21.146502145 +0000 UTC m=+1003.230102582" observedRunningTime="2025-09-30 14:36:22.407705745 +0000 UTC m=+1004.491306182" watchObservedRunningTime="2025-09-30 14:36:22.412479365 +0000 UTC m=+1004.496079782" Sep 30 14:36:22 crc kubenswrapper[4799]: I0930 14:36:22.435464 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" podStartSLOduration=5.792184185 podStartE2EDuration="33.435442016s" podCreationTimestamp="2025-09-30 14:35:49 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.40320297 +0000 UTC m=+975.486803397" lastFinishedPulling="2025-09-30 14:36:21.046460801 +0000 UTC m=+1003.130061228" observedRunningTime="2025-09-30 14:36:22.431283695 +0000 UTC m=+1004.514884122" watchObservedRunningTime="2025-09-30 14:36:22.435442016 +0000 UTC m=+1004.519042433" Sep 30 14:36:23 crc kubenswrapper[4799]: I0930 14:36:23.385439 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" Sep 30 14:36:23 crc kubenswrapper[4799]: I0930 14:36:23.386722 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-vq46v" Sep 30 14:36:23 crc kubenswrapper[4799]: I0930 14:36:23.388691 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-pxd4j" Sep 30 14:36:23 crc kubenswrapper[4799]: I0930 14:36:23.389037 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-c7c776c96-vwjtl" Sep 30 14:36:23 crc kubenswrapper[4799]: I0930 14:36:23.389836 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-2jnc8" Sep 30 14:36:23 crc kubenswrapper[4799]: I0930 14:36:23.389881 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-j7tgf" Sep 30 14:36:23 crc kubenswrapper[4799]: I0930 14:36:23.389925 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-9mwxx" Sep 30 14:36:23 crc kubenswrapper[4799]: I0930 14:36:23.392494 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7d857cc749-rzjpm" Sep 30 14:36:30 crc kubenswrapper[4799]: I0930 14:36:30.311814 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-8vxq2" Sep 30 14:36:30 crc kubenswrapper[4799]: I0930 14:36:30.433511 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-7975b88857-p75vq" Sep 30 14:36:30 crc kubenswrapper[4799]: I0930 14:36:30.645407 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-k955x" Sep 30 14:36:30 crc kubenswrapper[4799]: I0930 14:36:30.872211 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-qrwt4" Sep 30 14:36:31 crc kubenswrapper[4799]: I0930 14:36:31.218804 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-sz2rp" Sep 30 14:36:31 crc kubenswrapper[4799]: I0930 14:36:31.334351 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-f66b554c6-nxbv9" Sep 30 14:36:31 crc kubenswrapper[4799]: I0930 14:36:31.412489 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-76669f99c-c7mtd" Sep 30 14:36:31 crc kubenswrapper[4799]: I0930 14:36:31.505464 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:36:32 crc kubenswrapper[4799]: I0930 14:36:32.471219 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" event={"ID":"68487038-29af-45ea-8619-acd75f3d9a7d","Type":"ContainerStarted","Data":"fee2686563d6e98beb627410627de3b7cc98113ab4c94370529276c4677aba05"} Sep 30 14:36:32 crc kubenswrapper[4799]: I0930 14:36:32.472821 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:36:32 crc kubenswrapper[4799]: I0930 14:36:32.476598 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" event={"ID":"78d18534-fd0c-454f-80f3-b65ddd5a24c9","Type":"ContainerStarted","Data":"7497fbc69873918e4c020f57ff026efdfd3051f76b3cb6bc0404f1f7131653ba"} Sep 30 14:36:32 crc kubenswrapper[4799]: I0930 14:36:32.480355 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" event={"ID":"e0002170-fb28-4c87-a970-350f92c891bd","Type":"ContainerStarted","Data":"2fea52672cca90cf305840ec535e9ad29f76abc1d32fde47edcbe717e99a77a4"} Sep 30 14:36:32 crc kubenswrapper[4799]: I0930 14:36:32.481602 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" Sep 30 14:36:32 crc kubenswrapper[4799]: I0930 14:36:32.562950 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" podStartSLOduration=4.090957836 podStartE2EDuration="42.562920606s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.636456459 +0000 UTC m=+975.720056886" lastFinishedPulling="2025-09-30 14:36:32.108419229 +0000 UTC m=+1014.192019656" observedRunningTime="2025-09-30 14:36:32.527072598 +0000 UTC m=+1014.610673035" watchObservedRunningTime="2025-09-30 14:36:32.562920606 +0000 UTC m=+1014.646521033" Sep 30 14:36:32 crc kubenswrapper[4799]: I0930 14:36:32.573944 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-7tlqx" podStartSLOduration=2.835542998 podStartE2EDuration="41.573914647s" podCreationTimestamp="2025-09-30 14:35:51 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.37345043 +0000 UTC m=+975.457050857" lastFinishedPulling="2025-09-30 14:36:32.111822079 +0000 UTC m=+1014.195422506" observedRunningTime="2025-09-30 14:36:32.557830157 +0000 UTC m=+1014.641430594" watchObservedRunningTime="2025-09-30 14:36:32.573914647 +0000 UTC m=+1014.657515074" Sep 30 14:36:32 crc kubenswrapper[4799]: I0930 14:36:32.594733 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" podStartSLOduration=4.099824651 podStartE2EDuration="43.594706635s" podCreationTimestamp="2025-09-30 14:35:49 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.612761072 +0000 UTC m=+974.696361499" lastFinishedPulling="2025-09-30 14:36:32.107643056 +0000 UTC m=+1014.191243483" observedRunningTime="2025-09-30 14:36:32.584371143 +0000 UTC m=+1014.667971600" watchObservedRunningTime="2025-09-30 14:36:32.594706635 +0000 UTC m=+1014.678307072" Sep 30 14:36:34 crc kubenswrapper[4799]: I0930 14:36:34.498521 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" event={"ID":"f3e1a985-b655-4dc8-8e5f-f49a6459f512","Type":"ContainerStarted","Data":"be43fcb7d184df1a9ce5757821f08e5f9c96e7646ca91f6c269824080a2a35b4"} Sep 30 14:36:34 crc kubenswrapper[4799]: I0930 14:36:34.499379 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" Sep 30 14:36:34 crc kubenswrapper[4799]: I0930 14:36:34.526091 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" podStartSLOduration=2.901128321 podStartE2EDuration="44.526066977s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.613840183 +0000 UTC m=+974.697440610" lastFinishedPulling="2025-09-30 14:36:34.238778829 +0000 UTC m=+1016.322379266" observedRunningTime="2025-09-30 14:36:34.522035979 +0000 UTC m=+1016.605636406" watchObservedRunningTime="2025-09-30 14:36:34.526066977 +0000 UTC m=+1016.609667404" Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.509558 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" event={"ID":"8bc24269-6185-41a5-861b-19c88576c223","Type":"ContainerStarted","Data":"0dc7176fe99d64c1dc3c1f7e226805cf5819f8e529cb78bf8c99cdf395ccc29f"} Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.510340 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.511759 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" event={"ID":"34ffd2cc-7377-4160-b03d-cae022be0262","Type":"ContainerStarted","Data":"31ec46f39927c38aa24a8a60edfca98e426c6baefdf721ad9343e3e31b2201b3"} Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.511987 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.515043 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" event={"ID":"19539cfa-6f83-4636-8193-d2373038f353","Type":"ContainerStarted","Data":"b21f15d9566300b024479a33ac35f45a31dbd1751772559b74421c0568f75f25"} Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.515259 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.537170 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" podStartSLOduration=3.151035054 podStartE2EDuration="46.537043251s" podCreationTimestamp="2025-09-30 14:35:49 +0000 UTC" firstStartedPulling="2025-09-30 14:35:51.6384847 +0000 UTC m=+973.722085127" lastFinishedPulling="2025-09-30 14:36:35.024492897 +0000 UTC m=+1017.108093324" observedRunningTime="2025-09-30 14:36:35.533024094 +0000 UTC m=+1017.616624521" watchObservedRunningTime="2025-09-30 14:36:35.537043251 +0000 UTC m=+1017.620643678" Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.559979 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" podStartSLOduration=3.222883654 podStartE2EDuration="46.559956481s" podCreationTimestamp="2025-09-30 14:35:49 +0000 UTC" firstStartedPulling="2025-09-30 14:35:51.98946427 +0000 UTC m=+974.073064697" lastFinishedPulling="2025-09-30 14:36:35.326537097 +0000 UTC m=+1017.410137524" observedRunningTime="2025-09-30 14:36:35.555887822 +0000 UTC m=+1017.639488259" watchObservedRunningTime="2025-09-30 14:36:35.559956481 +0000 UTC m=+1017.643556908" Sep 30 14:36:35 crc kubenswrapper[4799]: I0930 14:36:35.577341 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" podStartSLOduration=3.434063521 podStartE2EDuration="45.577325819s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:52.881270611 +0000 UTC m=+974.964871038" lastFinishedPulling="2025-09-30 14:36:35.024532909 +0000 UTC m=+1017.108133336" observedRunningTime="2025-09-30 14:36:35.576403462 +0000 UTC m=+1017.660003889" watchObservedRunningTime="2025-09-30 14:36:35.577325819 +0000 UTC m=+1017.660926236" Sep 30 14:36:36 crc kubenswrapper[4799]: I0930 14:36:36.526036 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" event={"ID":"208f263c-4523-45e3-bf7e-19c8407bc599","Type":"ContainerStarted","Data":"12672569591a122acb26b3301307199b9a447277c08ad2021dd14d1cb1c4f6f3"} Sep 30 14:36:36 crc kubenswrapper[4799]: I0930 14:36:36.527267 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" Sep 30 14:36:36 crc kubenswrapper[4799]: I0930 14:36:36.552194 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" podStartSLOduration=4.003992404 podStartE2EDuration="46.552166878s" podCreationTimestamp="2025-09-30 14:35:50 +0000 UTC" firstStartedPulling="2025-09-30 14:35:53.378656962 +0000 UTC m=+975.462257389" lastFinishedPulling="2025-09-30 14:36:35.926831426 +0000 UTC m=+1018.010431863" observedRunningTime="2025-09-30 14:36:36.546138121 +0000 UTC m=+1018.629738568" watchObservedRunningTime="2025-09-30 14:36:36.552166878 +0000 UTC m=+1018.635767305" Sep 30 14:36:40 crc kubenswrapper[4799]: I0930 14:36:40.173374 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-mplqn" Sep 30 14:36:40 crc kubenswrapper[4799]: I0930 14:36:40.248553 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-gmxsp" Sep 30 14:36:40 crc kubenswrapper[4799]: I0930 14:36:40.446718 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-kqfcp" Sep 30 14:36:40 crc kubenswrapper[4799]: I0930 14:36:40.690332 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64d7b59854-9mvns" Sep 30 14:36:40 crc kubenswrapper[4799]: I0930 14:36:40.801033 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-76fcc6dc7c-rpsrk" Sep 30 14:36:41 crc kubenswrapper[4799]: I0930 14:36:41.254055 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-bc7dc7bd9-68qzx" Sep 30 14:36:42 crc kubenswrapper[4799]: I0930 14:36:42.411070 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6d776955-q7tzd" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.088630 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9sk5p"] Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.096457 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.103965 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.104805 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.110732 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9sk5p"] Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.111440 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.115468 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-mh9nm" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.195142 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m7pvk"] Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.196713 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.197487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kjj2\" (UniqueName: \"kubernetes.io/projected/df4d1b72-8dc1-4601-9766-1544375fb67e-kube-api-access-2kjj2\") pod \"dnsmasq-dns-675f4bcbfc-9sk5p\" (UID: \"df4d1b72-8dc1-4601-9766-1544375fb67e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.197606 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df4d1b72-8dc1-4601-9766-1544375fb67e-config\") pod \"dnsmasq-dns-675f4bcbfc-9sk5p\" (UID: \"df4d1b72-8dc1-4601-9766-1544375fb67e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.200483 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.220412 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m7pvk"] Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.299627 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-config\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.299711 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kjj2\" (UniqueName: \"kubernetes.io/projected/df4d1b72-8dc1-4601-9766-1544375fb67e-kube-api-access-2kjj2\") pod \"dnsmasq-dns-675f4bcbfc-9sk5p\" (UID: \"df4d1b72-8dc1-4601-9766-1544375fb67e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.299770 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2f92\" (UniqueName: \"kubernetes.io/projected/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-kube-api-access-s2f92\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.299814 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df4d1b72-8dc1-4601-9766-1544375fb67e-config\") pod \"dnsmasq-dns-675f4bcbfc-9sk5p\" (UID: \"df4d1b72-8dc1-4601-9766-1544375fb67e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.299836 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.301490 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df4d1b72-8dc1-4601-9766-1544375fb67e-config\") pod \"dnsmasq-dns-675f4bcbfc-9sk5p\" (UID: \"df4d1b72-8dc1-4601-9766-1544375fb67e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.330924 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kjj2\" (UniqueName: \"kubernetes.io/projected/df4d1b72-8dc1-4601-9766-1544375fb67e-kube-api-access-2kjj2\") pod \"dnsmasq-dns-675f4bcbfc-9sk5p\" (UID: \"df4d1b72-8dc1-4601-9766-1544375fb67e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.401307 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-config\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.401412 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2f92\" (UniqueName: \"kubernetes.io/projected/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-kube-api-access-s2f92\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.401458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.402844 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.403005 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-config\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.426059 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2f92\" (UniqueName: \"kubernetes.io/projected/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-kube-api-access-s2f92\") pod \"dnsmasq-dns-78dd6ddcc-m7pvk\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.426413 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.517497 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:00 crc kubenswrapper[4799]: I0930 14:37:00.971606 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9sk5p"] Sep 30 14:37:01 crc kubenswrapper[4799]: I0930 14:37:01.100262 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m7pvk"] Sep 30 14:37:01 crc kubenswrapper[4799]: W0930 14:37:01.110587 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1709326_a1b4_45ad_bb14_f3d0b2eb5124.slice/crio-b2e64ad5f8f7802720dcba91907c74446423b38395f66fc3e84db7c314299dcc WatchSource:0}: Error finding container b2e64ad5f8f7802720dcba91907c74446423b38395f66fc3e84db7c314299dcc: Status 404 returned error can't find the container with id b2e64ad5f8f7802720dcba91907c74446423b38395f66fc3e84db7c314299dcc Sep 30 14:37:01 crc kubenswrapper[4799]: I0930 14:37:01.748597 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" event={"ID":"df4d1b72-8dc1-4601-9766-1544375fb67e","Type":"ContainerStarted","Data":"b335795175d025ca21c2fbeb695fbaf8fc42704a06bf366b3b3c7fd8d21f2957"} Sep 30 14:37:01 crc kubenswrapper[4799]: I0930 14:37:01.751298 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" event={"ID":"e1709326-a1b4-45ad-bb14-f3d0b2eb5124","Type":"ContainerStarted","Data":"b2e64ad5f8f7802720dcba91907c74446423b38395f66fc3e84db7c314299dcc"} Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.123373 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9sk5p"] Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.191547 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-l6mn9"] Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.197198 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.255426 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-l6mn9"] Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.275184 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-config\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.275257 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnv4m\" (UniqueName: \"kubernetes.io/projected/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-kube-api-access-dnv4m\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.275310 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.377949 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.378075 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-config\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.378112 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnv4m\" (UniqueName: \"kubernetes.io/projected/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-kube-api-access-dnv4m\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.380778 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.389940 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-config\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.416296 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnv4m\" (UniqueName: \"kubernetes.io/projected/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-kube-api-access-dnv4m\") pod \"dnsmasq-dns-5ccc8479f9-l6mn9\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.544587 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.549498 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m7pvk"] Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.590847 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rctd8"] Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.597903 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.618860 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rctd8"] Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.683815 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-config\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.683902 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrqr4\" (UniqueName: \"kubernetes.io/projected/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-kube-api-access-rrqr4\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.683987 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.790154 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrqr4\" (UniqueName: \"kubernetes.io/projected/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-kube-api-access-rrqr4\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.790281 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.790330 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-config\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.791365 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-config\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.792507 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.842554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrqr4\" (UniqueName: \"kubernetes.io/projected/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-kube-api-access-rrqr4\") pod \"dnsmasq-dns-57d769cc4f-rctd8\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:03 crc kubenswrapper[4799]: I0930 14:37:03.973204 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.344317 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.347826 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.352412 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.352714 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.355025 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.356708 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.356818 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.356873 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g4bn4" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.357051 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.363332 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423320 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423388 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423424 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423467 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423520 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef6dcf7e-70a0-4759-afef-53e70b800ba0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423563 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef6dcf7e-70a0-4759-afef-53e70b800ba0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423590 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423620 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz59q\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-kube-api-access-gz59q\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423710 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.423780 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.460184 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-l6mn9"] Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525135 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525209 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525238 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525276 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525351 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525377 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525405 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef6dcf7e-70a0-4759-afef-53e70b800ba0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525431 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef6dcf7e-70a0-4759-afef-53e70b800ba0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525468 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525499 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz59q\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-kube-api-access-gz59q\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.525552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.526704 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.528110 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.528477 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.529252 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.530759 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.531319 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.537115 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef6dcf7e-70a0-4759-afef-53e70b800ba0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.537147 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.546914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef6dcf7e-70a0-4759-afef-53e70b800ba0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.547419 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.557268 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz59q\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-kube-api-access-gz59q\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.573575 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.679846 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rctd8"] Sep 30 14:37:04 crc kubenswrapper[4799]: W0930 14:37:04.698293 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddde499c4_cff7_41b0_80b2_2a9ee3f7b0df.slice/crio-fdf8b194210af49b0144357c37680d05771954f1edb94c4a0201851b7e2d3ff0 WatchSource:0}: Error finding container fdf8b194210af49b0144357c37680d05771954f1edb94c4a0201851b7e2d3ff0: Status 404 returned error can't find the container with id fdf8b194210af49b0144357c37680d05771954f1edb94c4a0201851b7e2d3ff0 Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.722287 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.790574 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.803319 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.803424 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.821904 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.822126 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.822159 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.822394 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.824359 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.826423 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-qd87k" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.830506 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.848617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" event={"ID":"4adfc365-32d8-44a1-9eca-11eabcdc1ab6","Type":"ContainerStarted","Data":"aefffeddb19eb23a595487b08521d8cb65060aa71d463ef6d1cc12eccff8ace9"} Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.851078 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" event={"ID":"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df","Type":"ContainerStarted","Data":"fdf8b194210af49b0144357c37680d05771954f1edb94c4a0201851b7e2d3ff0"} Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935435 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jb4r\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-kube-api-access-7jb4r\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935568 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935666 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-config-data\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935691 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935730 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935804 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935869 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935894 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.935963 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bb82f13-0450-42a8-9c8a-0c174c534382-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.936360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:04 crc kubenswrapper[4799]: I0930 14:37:04.936472 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bb82f13-0450-42a8-9c8a-0c174c534382-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.039708 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042066 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042122 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042204 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bb82f13-0450-42a8-9c8a-0c174c534382-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042286 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bb82f13-0450-42a8-9c8a-0c174c534382-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jb4r\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-kube-api-access-7jb4r\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042507 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042860 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-config-data\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042909 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.043050 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.043528 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.044952 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.045063 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-config-data\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.042696 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.045334 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.045597 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.053335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bb82f13-0450-42a8-9c8a-0c174c534382-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.059875 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.068666 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.070817 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jb4r\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-kube-api-access-7jb4r\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.071902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bb82f13-0450-42a8-9c8a-0c174c534382-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.095475 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.132104 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.408413 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.801681 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:37:05 crc kubenswrapper[4799]: W0930 14:37:05.815991 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bb82f13_0450_42a8_9c8a_0c174c534382.slice/crio-0e93bb0e81164ea28f99c0b1bd326238744ae6c6f285ce255c65c499dd713c96 WatchSource:0}: Error finding container 0e93bb0e81164ea28f99c0b1bd326238744ae6c6f285ce255c65c499dd713c96: Status 404 returned error can't find the container with id 0e93bb0e81164ea28f99c0b1bd326238744ae6c6f285ce255c65c499dd713c96 Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.893624 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef6dcf7e-70a0-4759-afef-53e70b800ba0","Type":"ContainerStarted","Data":"4a3f9ce55d0a4a85e0c47fcefbcac947f37982c69748a676af531e564ccf0152"} Sep 30 14:37:05 crc kubenswrapper[4799]: I0930 14:37:05.902997 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb82f13-0450-42a8-9c8a-0c174c534382","Type":"ContainerStarted","Data":"0e93bb0e81164ea28f99c0b1bd326238744ae6c6f285ce255c65c499dd713c96"} Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.407967 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.411611 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.416178 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.416532 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.431334 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.431872 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-n6pd6" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.435280 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.435550 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.437942 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.472103 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.482522 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.490614 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.491537 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.491719 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.491961 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-v8rh5" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.531778 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638602 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638678 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b86436d8-96e2-44aa-9b4c-b9292a1ca129-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638707 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638728 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638764 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638792 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/33a2e46f-eaef-4d54-8af4-ef388d3425dc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638815 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-kolla-config\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638834 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638857 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz8vl\" (UniqueName: \"kubernetes.io/projected/33a2e46f-eaef-4d54-8af4-ef388d3425dc-kube-api-access-sz8vl\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638876 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638896 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638915 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638960 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztw7d\" (UniqueName: \"kubernetes.io/projected/b86436d8-96e2-44aa-9b4c-b9292a1ca129-kube-api-access-ztw7d\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.638983 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.639009 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.639030 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.639057 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-secrets\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.639079 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-config-data-default\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741004 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741071 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-secrets\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741093 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-config-data-default\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741115 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741137 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b86436d8-96e2-44aa-9b4c-b9292a1ca129-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741154 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741204 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741242 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/33a2e46f-eaef-4d54-8af4-ef388d3425dc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741263 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-kolla-config\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741287 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741313 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz8vl\" (UniqueName: \"kubernetes.io/projected/33a2e46f-eaef-4d54-8af4-ef388d3425dc-kube-api-access-sz8vl\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741336 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741368 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741411 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztw7d\" (UniqueName: \"kubernetes.io/projected/b86436d8-96e2-44aa-9b4c-b9292a1ca129-kube-api-access-ztw7d\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741466 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.741811 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.742925 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/33a2e46f-eaef-4d54-8af4-ef388d3425dc-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.743183 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.744496 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-config-data-default\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.745024 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.745349 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b86436d8-96e2-44aa-9b4c-b9292a1ca129-kolla-config\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.745571 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.759309 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b86436d8-96e2-44aa-9b4c-b9292a1ca129-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.760238 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.761069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/33a2e46f-eaef-4d54-8af4-ef388d3425dc-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.773389 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.774809 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.775318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-secrets\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.778702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.797086 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b86436d8-96e2-44aa-9b4c-b9292a1ca129-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.801221 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a2e46f-eaef-4d54-8af4-ef388d3425dc-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.803062 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztw7d\" (UniqueName: \"kubernetes.io/projected/b86436d8-96e2-44aa-9b4c-b9292a1ca129-kube-api-access-ztw7d\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.824281 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz8vl\" (UniqueName: \"kubernetes.io/projected/33a2e46f-eaef-4d54-8af4-ef388d3425dc-kube-api-access-sz8vl\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.824668 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"b86436d8-96e2-44aa-9b4c-b9292a1ca129\") " pod="openstack/openstack-galera-0" Sep 30 14:37:07 crc kubenswrapper[4799]: I0930 14:37:07.895978 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"33a2e46f-eaef-4d54-8af4-ef388d3425dc\") " pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.060717 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.143527 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.392530 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.394114 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.400269 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-h78nf" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.400536 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.416914 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.445535 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.476224 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.476581 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-config-data\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.476686 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-kolla-config\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.476795 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgbp\" (UniqueName: \"kubernetes.io/projected/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-kube-api-access-2xgbp\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.476916 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.593923 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.594845 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-config-data\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.596851 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-kolla-config\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.597117 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgbp\" (UniqueName: \"kubernetes.io/projected/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-kube-api-access-2xgbp\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.597436 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.615999 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.616673 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.619131 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-kolla-config\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.622492 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-config-data\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.672229 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgbp\" (UniqueName: \"kubernetes.io/projected/e3bd7a13-1fa3-4901-aeb6-3b9011ae330b-kube-api-access-2xgbp\") pod \"memcached-0\" (UID: \"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b\") " pod="openstack/memcached-0" Sep 30 14:37:08 crc kubenswrapper[4799]: I0930 14:37:08.750166 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Sep 30 14:37:09 crc kubenswrapper[4799]: I0930 14:37:09.618403 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Sep 30 14:37:09 crc kubenswrapper[4799]: I0930 14:37:09.768040 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Sep 30 14:37:09 crc kubenswrapper[4799]: W0930 14:37:09.877498 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb86436d8_96e2_44aa_9b4c_b9292a1ca129.slice/crio-7ca643f135fc398061e5b4f38705a309824080336bb33ca29f98cef0609bd75d WatchSource:0}: Error finding container 7ca643f135fc398061e5b4f38705a309824080336bb33ca29f98cef0609bd75d: Status 404 returned error can't find the container with id 7ca643f135fc398061e5b4f38705a309824080336bb33ca29f98cef0609bd75d Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.168982 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.194404 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b86436d8-96e2-44aa-9b4c-b9292a1ca129","Type":"ContainerStarted","Data":"7ca643f135fc398061e5b4f38705a309824080336bb33ca29f98cef0609bd75d"} Sep 30 14:37:10 crc kubenswrapper[4799]: W0930 14:37:10.198354 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3bd7a13_1fa3_4901_aeb6_3b9011ae330b.slice/crio-c13ad2148a96ade1f45dc24e2f34cfcc840a8affd9b2a2916bee3780b61d4bfd WatchSource:0}: Error finding container c13ad2148a96ade1f45dc24e2f34cfcc840a8affd9b2a2916bee3780b61d4bfd: Status 404 returned error can't find the container with id c13ad2148a96ade1f45dc24e2f34cfcc840a8affd9b2a2916bee3780b61d4bfd Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.200099 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"33a2e46f-eaef-4d54-8af4-ef388d3425dc","Type":"ContainerStarted","Data":"3ac37305216ed821a69e57757723a014b5efa052a9ea43d39d5816bd07854b9b"} Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.611917 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.627633 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.632327 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-p5h9m" Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.644646 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.681451 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9g7f\" (UniqueName: \"kubernetes.io/projected/ae8631cb-f4a0-4d95-afb4-64129fc0475a-kube-api-access-g9g7f\") pod \"kube-state-metrics-0\" (UID: \"ae8631cb-f4a0-4d95-afb4-64129fc0475a\") " pod="openstack/kube-state-metrics-0" Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.785097 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9g7f\" (UniqueName: \"kubernetes.io/projected/ae8631cb-f4a0-4d95-afb4-64129fc0475a-kube-api-access-g9g7f\") pod \"kube-state-metrics-0\" (UID: \"ae8631cb-f4a0-4d95-afb4-64129fc0475a\") " pod="openstack/kube-state-metrics-0" Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.827831 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9g7f\" (UniqueName: \"kubernetes.io/projected/ae8631cb-f4a0-4d95-afb4-64129fc0475a-kube-api-access-g9g7f\") pod \"kube-state-metrics-0\" (UID: \"ae8631cb-f4a0-4d95-afb4-64129fc0475a\") " pod="openstack/kube-state-metrics-0" Sep 30 14:37:10 crc kubenswrapper[4799]: I0930 14:37:10.961637 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:37:11 crc kubenswrapper[4799]: I0930 14:37:11.219429 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b","Type":"ContainerStarted","Data":"c13ad2148a96ade1f45dc24e2f34cfcc840a8affd9b2a2916bee3780b61d4bfd"} Sep 30 14:37:11 crc kubenswrapper[4799]: I0930 14:37:11.862365 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.144920 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2kvml"] Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.146186 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.154005 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-v6skl" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.154264 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.154386 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.177786 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-v2rs4"] Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.179778 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.184634 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2kvml"] Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.229030 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v2rs4"] Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.233101 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-run\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.233298 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-lib\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.233358 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-etc-ovs\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.233414 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddxp4\" (UniqueName: \"kubernetes.io/projected/17bda665-adad-40d3-9465-46013020af4e-kube-api-access-ddxp4\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.233481 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17bda665-adad-40d3-9465-46013020af4e-scripts\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.233583 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-log\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.284006 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ae8631cb-f4a0-4d95-afb4-64129fc0475a","Type":"ContainerStarted","Data":"4c288664139750daa91b7f91c69e7f62e450cc61dada8c6ec3718c7a76d19dd6"} Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.335947 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c25ft\" (UniqueName: \"kubernetes.io/projected/cbefa245-5ad5-4303-b166-59574e8a0c65-kube-api-access-c25ft\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.336055 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbefa245-5ad5-4303-b166-59574e8a0c65-scripts\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.336087 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-run-ovn\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.336125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddxp4\" (UniqueName: \"kubernetes.io/projected/17bda665-adad-40d3-9465-46013020af4e-kube-api-access-ddxp4\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.336829 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbefa245-5ad5-4303-b166-59574e8a0c65-ovn-controller-tls-certs\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.336891 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17bda665-adad-40d3-9465-46013020af4e-scripts\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.337140 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-log\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.337262 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-run\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.337298 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-run\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.337389 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-lib\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.337426 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-log-ovn\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.337453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbefa245-5ad5-4303-b166-59574e8a0c65-combined-ca-bundle\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.337478 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-etc-ovs\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.338166 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-log\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.338241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-run\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.338394 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-var-lib\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.338456 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/17bda665-adad-40d3-9465-46013020af4e-etc-ovs\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.345950 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17bda665-adad-40d3-9465-46013020af4e-scripts\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.357467 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddxp4\" (UniqueName: \"kubernetes.io/projected/17bda665-adad-40d3-9465-46013020af4e-kube-api-access-ddxp4\") pod \"ovn-controller-ovs-v2rs4\" (UID: \"17bda665-adad-40d3-9465-46013020af4e\") " pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.438675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-run\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.438767 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-log-ovn\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.438799 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbefa245-5ad5-4303-b166-59574e8a0c65-combined-ca-bundle\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.438834 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c25ft\" (UniqueName: \"kubernetes.io/projected/cbefa245-5ad5-4303-b166-59574e8a0c65-kube-api-access-c25ft\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.438865 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbefa245-5ad5-4303-b166-59574e8a0c65-scripts\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.438887 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-run-ovn\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.438934 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbefa245-5ad5-4303-b166-59574e8a0c65-ovn-controller-tls-certs\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.444638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbefa245-5ad5-4303-b166-59574e8a0c65-combined-ca-bundle\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.444964 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-log-ovn\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.444909 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-run\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.445156 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cbefa245-5ad5-4303-b166-59574e8a0c65-var-run-ovn\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.448084 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbefa245-5ad5-4303-b166-59574e8a0c65-scripts\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.459701 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbefa245-5ad5-4303-b166-59574e8a0c65-ovn-controller-tls-certs\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.488356 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c25ft\" (UniqueName: \"kubernetes.io/projected/cbefa245-5ad5-4303-b166-59574e8a0c65-kube-api-access-c25ft\") pod \"ovn-controller-2kvml\" (UID: \"cbefa245-5ad5-4303-b166-59574e8a0c65\") " pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.495427 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2kvml" Sep 30 14:37:12 crc kubenswrapper[4799]: I0930 14:37:12.536468 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:37:13 crc kubenswrapper[4799]: I0930 14:37:13.502215 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2kvml"] Sep 30 14:37:13 crc kubenswrapper[4799]: W0930 14:37:13.581253 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbefa245_5ad5_4303_b166_59574e8a0c65.slice/crio-24fc3f91c0cc009ba198484095287ef78c547fa7d989484116b91f431d713c65 WatchSource:0}: Error finding container 24fc3f91c0cc009ba198484095287ef78c547fa7d989484116b91f431d713c65: Status 404 returned error can't find the container with id 24fc3f91c0cc009ba198484095287ef78c547fa7d989484116b91f431d713c65 Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.176642 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.179534 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.183340 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.183760 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.183907 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.184113 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.184557 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-zjzzx" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.200254 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.288015 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-config\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.288106 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.288128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mftqz\" (UniqueName: \"kubernetes.io/projected/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-kube-api-access-mftqz\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.288145 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.288196 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.288217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.288241 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.288309 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.322363 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2kvml" event={"ID":"cbefa245-5ad5-4303-b166-59574e8a0c65","Type":"ContainerStarted","Data":"24fc3f91c0cc009ba198484095287ef78c547fa7d989484116b91f431d713c65"} Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.391951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.392029 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.392072 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.392129 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.392163 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-config\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.392266 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.392301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mftqz\" (UniqueName: \"kubernetes.io/projected/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-kube-api-access-mftqz\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.392337 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.394017 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.394871 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.395258 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.400557 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-config\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.403171 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.406725 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.428375 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.444305 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mftqz\" (UniqueName: \"kubernetes.io/projected/c146ba83-511e-4ff2-9a3e-6c83c4c43e9e-kube-api-access-mftqz\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.462659 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e\") " pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:14 crc kubenswrapper[4799]: I0930 14:37:14.513617 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.184118 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v2rs4"] Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.240734 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-r82xm"] Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.242217 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.245037 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.264547 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-r82xm"] Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.290692 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9ee89b02-9e5e-4c62-ade3-c4a308feefde-ovs-rundir\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.290778 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9ee89b02-9e5e-4c62-ade3-c4a308feefde-ovn-rundir\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.290806 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee89b02-9e5e-4c62-ade3-c4a308feefde-combined-ca-bundle\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.290867 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ee89b02-9e5e-4c62-ade3-c4a308feefde-config\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.290927 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ee89b02-9e5e-4c62-ade3-c4a308feefde-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.290959 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr2b5\" (UniqueName: \"kubernetes.io/projected/9ee89b02-9e5e-4c62-ade3-c4a308feefde-kube-api-access-rr2b5\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.380124 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v2rs4" event={"ID":"17bda665-adad-40d3-9465-46013020af4e","Type":"ContainerStarted","Data":"d7d862dc9b9f7275f89f634cb5073f8d7dab6e9d445b0bf7d8bb45cb2bc4728b"} Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.392186 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9ee89b02-9e5e-4c62-ade3-c4a308feefde-ovs-rundir\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.392234 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9ee89b02-9e5e-4c62-ade3-c4a308feefde-ovn-rundir\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.392267 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee89b02-9e5e-4c62-ade3-c4a308feefde-combined-ca-bundle\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.392311 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ee89b02-9e5e-4c62-ade3-c4a308feefde-config\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.392356 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ee89b02-9e5e-4c62-ade3-c4a308feefde-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.392380 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr2b5\" (UniqueName: \"kubernetes.io/projected/9ee89b02-9e5e-4c62-ade3-c4a308feefde-kube-api-access-rr2b5\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.392698 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9ee89b02-9e5e-4c62-ade3-c4a308feefde-ovs-rundir\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.393906 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9ee89b02-9e5e-4c62-ade3-c4a308feefde-ovn-rundir\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.395671 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ee89b02-9e5e-4c62-ade3-c4a308feefde-config\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.404346 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee89b02-9e5e-4c62-ade3-c4a308feefde-combined-ca-bundle\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.426451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr2b5\" (UniqueName: \"kubernetes.io/projected/9ee89b02-9e5e-4c62-ade3-c4a308feefde-kube-api-access-rr2b5\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.445960 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ee89b02-9e5e-4c62-ade3-c4a308feefde-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r82xm\" (UID: \"9ee89b02-9e5e-4c62-ade3-c4a308feefde\") " pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.622641 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-r82xm" Sep 30 14:37:16 crc kubenswrapper[4799]: I0930 14:37:16.866388 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.323897 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.327297 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.335959 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.336138 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.336416 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-p9bjj" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.336814 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.344863 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.388760 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-r82xm"] Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.423143 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.423207 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2mz5\" (UniqueName: \"kubernetes.io/projected/8d8fe29c-5270-400a-9fc6-1024482dd54d-kube-api-access-p2mz5\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.423256 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.423285 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.423335 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d8fe29c-5270-400a-9fc6-1024482dd54d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.423381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8d8fe29c-5270-400a-9fc6-1024482dd54d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.423408 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d8fe29c-5270-400a-9fc6-1024482dd54d-config\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.423455 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.427389 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e","Type":"ContainerStarted","Data":"bfc501161aeb69107cf546d2bf332efa5f042e4913c84b1ee64cc4607f6d8822"} Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.525755 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d8fe29c-5270-400a-9fc6-1024482dd54d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.526391 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8d8fe29c-5270-400a-9fc6-1024482dd54d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.526443 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d8fe29c-5270-400a-9fc6-1024482dd54d-config\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.526499 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.526556 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.526596 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2mz5\" (UniqueName: \"kubernetes.io/projected/8d8fe29c-5270-400a-9fc6-1024482dd54d-kube-api-access-p2mz5\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.526639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.526706 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.527233 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8d8fe29c-5270-400a-9fc6-1024482dd54d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.529320 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d8fe29c-5270-400a-9fc6-1024482dd54d-config\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.529593 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.530397 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d8fe29c-5270-400a-9fc6-1024482dd54d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.535931 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.547031 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.548919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d8fe29c-5270-400a-9fc6-1024482dd54d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.551284 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2mz5\" (UniqueName: \"kubernetes.io/projected/8d8fe29c-5270-400a-9fc6-1024482dd54d-kube-api-access-p2mz5\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.562363 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8d8fe29c-5270-400a-9fc6-1024482dd54d\") " pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:17 crc kubenswrapper[4799]: I0930 14:37:17.669354 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Sep 30 14:37:18 crc kubenswrapper[4799]: I0930 14:37:18.447257 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-r82xm" event={"ID":"9ee89b02-9e5e-4c62-ade3-c4a308feefde","Type":"ContainerStarted","Data":"39b929568da603eb26a121808b4822cfd225d9643d4916c66647c87da8bf3606"} Sep 30 14:37:18 crc kubenswrapper[4799]: I0930 14:37:18.907430 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Sep 30 14:37:24 crc kubenswrapper[4799]: I0930 14:37:24.517916 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8d8fe29c-5270-400a-9fc6-1024482dd54d","Type":"ContainerStarted","Data":"5428364c9f74ff0b3b9e60349e76fd748a9475c468be946adad91bccfbb59afd"} Sep 30 14:37:29 crc kubenswrapper[4799]: I0930 14:37:29.649325 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:37:29 crc kubenswrapper[4799]: I0930 14:37:29.650291 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:37:31 crc kubenswrapper[4799]: E0930 14:37:31.768839 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Sep 30 14:37:31 crc kubenswrapper[4799]: E0930 14:37:31.769070 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ztw7d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(b86436d8-96e2-44aa-9b4c-b9292a1ca129): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:31 crc kubenswrapper[4799]: E0930 14:37:31.770284 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="b86436d8-96e2-44aa-9b4c-b9292a1ca129" Sep 30 14:37:31 crc kubenswrapper[4799]: E0930 14:37:31.872775 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Sep 30 14:37:31 crc kubenswrapper[4799]: E0930 14:37:31.872995 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sz8vl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(33a2e46f-eaef-4d54-8af4-ef388d3425dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:31 crc kubenswrapper[4799]: E0930 14:37:31.874236 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="33a2e46f-eaef-4d54-8af4-ef388d3425dc" Sep 30 14:37:32 crc kubenswrapper[4799]: E0930 14:37:32.594244 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="33a2e46f-eaef-4d54-8af4-ef388d3425dc" Sep 30 14:37:32 crc kubenswrapper[4799]: E0930 14:37:32.594784 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="b86436d8-96e2-44aa-9b4c-b9292a1ca129" Sep 30 14:37:40 crc kubenswrapper[4799]: E0930 14:37:40.200086 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage2740509720/2\": happened during read: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Sep 30 14:37:40 crc kubenswrapper[4799]: E0930 14:37:40.200852 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n685h8fh569h5dbh644h5fdh55bh5c6h695h659h8dh549h6hc9h67bh68bhbfh594h59bh655h5b6h596h5b7h5bch59h566h6h86h5c4h96h67dh5c8q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p2mz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(8d8fe29c-5270-400a-9fc6-1024482dd54d): ErrImagePull: rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage2740509720/2\": happened during read: context canceled" logger="UnhandledError" Sep 30 14:37:40 crc kubenswrapper[4799]: E0930 14:37:40.322389 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Sep 30 14:37:40 crc kubenswrapper[4799]: E0930 14:37:40.322674 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n579h594hc5h79h688h5c5h5f9h5cdh5f7hfch5f9h7ch646h5ddh559hch56dh5f7hd4h74h56bh8bh6bh679h55fh554h669hfbh56bh9dh85h675q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2xgbp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(e3bd7a13-1fa3-4901-aeb6-3b9011ae330b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:40 crc kubenswrapper[4799]: E0930 14:37:40.323933 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="e3bd7a13-1fa3-4901-aeb6-3b9011ae330b" Sep 30 14:37:40 crc kubenswrapper[4799]: E0930 14:37:40.659267 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="e3bd7a13-1fa3-4901-aeb6-3b9011ae330b" Sep 30 14:37:41 crc kubenswrapper[4799]: E0930 14:37:41.614034 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Sep 30 14:37:41 crc kubenswrapper[4799]: E0930 14:37:41.614282 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gz59q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(ef6dcf7e-70a0-4759-afef-53e70b800ba0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:41 crc kubenswrapper[4799]: E0930 14:37:41.615346 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" Sep 30 14:37:41 crc kubenswrapper[4799]: E0930 14:37:41.670841 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" Sep 30 14:37:42 crc kubenswrapper[4799]: E0930 14:37:42.866133 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Sep 30 14:37:42 crc kubenswrapper[4799]: E0930 14:37:42.866839 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7jb4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(1bb82f13-0450-42a8-9c8a-0c174c534382): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:42 crc kubenswrapper[4799]: E0930 14:37:42.868612 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" Sep 30 14:37:43 crc kubenswrapper[4799]: E0930 14:37:43.689924 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" Sep 30 14:37:43 crc kubenswrapper[4799]: E0930 14:37:43.932243 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified" Sep 30 14:37:43 crc kubenswrapper[4799]: E0930 14:37:43.932456 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:ovsdb-server-init,Image:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,Command:[/usr/local/bin/container-scripts/init-ovsdb-server.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n645h5d9h5f8h67chf6h647h65dh97h56bh5c6h64ch5dfh77h54dh65ch679h54fh576hdch98h5h67dh7ch64dh65bh599h697h5f7h5d6h689hcdh654q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-ovs,ReadOnly:false,MountPath:/etc/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log,ReadOnly:false,MountPath:/var/log/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-lib,ReadOnly:false,MountPath:/var/lib/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ddxp4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-ovs-v2rs4_openstack(17bda665-adad-40d3-9465-46013020af4e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:43 crc kubenswrapper[4799]: E0930 14:37:43.933697 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-ovs-v2rs4" podUID="17bda665-adad-40d3-9465-46013020af4e" Sep 30 14:37:44 crc kubenswrapper[4799]: E0930 14:37:44.690943 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified\\\"\"" pod="openstack/ovn-controller-ovs-v2rs4" podUID="17bda665-adad-40d3-9465-46013020af4e" Sep 30 14:37:44 crc kubenswrapper[4799]: E0930 14:37:44.882718 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 30 14:37:44 crc kubenswrapper[4799]: E0930 14:37:44.882987 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2f92,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-m7pvk_openstack(e1709326-a1b4-45ad-bb14-f3d0b2eb5124): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:44 crc kubenswrapper[4799]: E0930 14:37:44.884201 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" podUID="e1709326-a1b4-45ad-bb14-f3d0b2eb5124" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.026623 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.026838 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2kjj2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-9sk5p_openstack(df4d1b72-8dc1-4601-9766-1544375fb67e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.028064 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" podUID="df4d1b72-8dc1-4601-9766-1544375fb67e" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.077109 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.078224 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n645h5d9h5f8h67chf6h647h65dh97h56bh5c6h64ch5dfh77h54dh65ch679h54fh576hdch98h5h67dh7ch64dh65bh599h697h5f7h5d6h689hcdh654q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c25ft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-2kvml_openstack(cbefa245-5ad5-4303-b166-59574e8a0c65): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.079486 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-2kvml" podUID="cbefa245-5ad5-4303-b166-59574e8a0c65" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.112888 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.113124 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dnv4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-l6mn9_openstack(4adfc365-32d8-44a1-9eca-11eabcdc1ab6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.114247 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" podUID="4adfc365-32d8-44a1-9eca-11eabcdc1ab6" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.209904 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.210433 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rrqr4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-rctd8_openstack(dde499c4-cff7-41b0-80b2-2a9ee3f7b0df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.211632 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" podUID="dde499c4-cff7-41b0-80b2-2a9ee3f7b0df" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.701224 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-2kvml" podUID="cbefa245-5ad5-4303-b166-59574e8a0c65" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.702713 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" podUID="dde499c4-cff7-41b0-80b2-2a9ee3f7b0df" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.702769 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" podUID="4adfc365-32d8-44a1-9eca-11eabcdc1ab6" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.974414 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Sep 30 14:37:45 crc kubenswrapper[4799]: E0930 14:37:45.974629 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb8h57hcbhd8h5bdh58h58dhf7hcbhfh5bfh5dhc8h65dh56bh55dh7dhb5h9fh69h658hcdh64h57chb9h5b5h655hfch5ch67bh5bch695q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mftqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(c146ba83-511e-4ff2-9a3e-6c83c4c43e9e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:46 crc kubenswrapper[4799]: I0930 14:37:46.814795 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:46 crc kubenswrapper[4799]: I0930 14:37:46.911601 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-dns-svc\") pod \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " Sep 30 14:37:46 crc kubenswrapper[4799]: I0930 14:37:46.911721 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-config\") pod \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " Sep 30 14:37:46 crc kubenswrapper[4799]: I0930 14:37:46.911793 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2f92\" (UniqueName: \"kubernetes.io/projected/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-kube-api-access-s2f92\") pod \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\" (UID: \"e1709326-a1b4-45ad-bb14-f3d0b2eb5124\") " Sep 30 14:37:46 crc kubenswrapper[4799]: I0930 14:37:46.912525 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e1709326-a1b4-45ad-bb14-f3d0b2eb5124" (UID: "e1709326-a1b4-45ad-bb14-f3d0b2eb5124"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:37:46 crc kubenswrapper[4799]: I0930 14:37:46.912540 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-config" (OuterVolumeSpecName: "config") pod "e1709326-a1b4-45ad-bb14-f3d0b2eb5124" (UID: "e1709326-a1b4-45ad-bb14-f3d0b2eb5124"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:37:46 crc kubenswrapper[4799]: I0930 14:37:46.918752 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-kube-api-access-s2f92" (OuterVolumeSpecName: "kube-api-access-s2f92") pod "e1709326-a1b4-45ad-bb14-f3d0b2eb5124" (UID: "e1709326-a1b4-45ad-bb14-f3d0b2eb5124"). InnerVolumeSpecName "kube-api-access-s2f92". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.014218 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.014559 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.014623 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2f92\" (UniqueName: \"kubernetes.io/projected/e1709326-a1b4-45ad-bb14-f3d0b2eb5124-kube-api-access-s2f92\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:47 crc kubenswrapper[4799]: E0930 14:37:47.250719 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Sep 30 14:37:47 crc kubenswrapper[4799]: E0930 14:37:47.250899 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovs-rundir,ReadOnly:true,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-rundir,ReadOnly:true,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rr2b5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-metrics-r82xm_openstack(9ee89b02-9e5e-4c62-ade3-c4a308feefde): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:37:47 crc kubenswrapper[4799]: E0930 14:37:47.252143 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-metrics-r82xm" podUID="9ee89b02-9e5e-4c62-ade3-c4a308feefde" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.307850 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.421238 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df4d1b72-8dc1-4601-9766-1544375fb67e-config\") pod \"df4d1b72-8dc1-4601-9766-1544375fb67e\" (UID: \"df4d1b72-8dc1-4601-9766-1544375fb67e\") " Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.421301 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kjj2\" (UniqueName: \"kubernetes.io/projected/df4d1b72-8dc1-4601-9766-1544375fb67e-kube-api-access-2kjj2\") pod \"df4d1b72-8dc1-4601-9766-1544375fb67e\" (UID: \"df4d1b72-8dc1-4601-9766-1544375fb67e\") " Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.422216 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df4d1b72-8dc1-4601-9766-1544375fb67e-config" (OuterVolumeSpecName: "config") pod "df4d1b72-8dc1-4601-9766-1544375fb67e" (UID: "df4d1b72-8dc1-4601-9766-1544375fb67e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.425507 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df4d1b72-8dc1-4601-9766-1544375fb67e-kube-api-access-2kjj2" (OuterVolumeSpecName: "kube-api-access-2kjj2") pod "df4d1b72-8dc1-4601-9766-1544375fb67e" (UID: "df4d1b72-8dc1-4601-9766-1544375fb67e"). InnerVolumeSpecName "kube-api-access-2kjj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.526279 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df4d1b72-8dc1-4601-9766-1544375fb67e-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.526316 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kjj2\" (UniqueName: \"kubernetes.io/projected/df4d1b72-8dc1-4601-9766-1544375fb67e-kube-api-access-2kjj2\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.716965 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" event={"ID":"e1709326-a1b4-45ad-bb14-f3d0b2eb5124","Type":"ContainerDied","Data":"b2e64ad5f8f7802720dcba91907c74446423b38395f66fc3e84db7c314299dcc"} Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.717024 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-m7pvk" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.722408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" event={"ID":"df4d1b72-8dc1-4601-9766-1544375fb67e","Type":"ContainerDied","Data":"b335795175d025ca21c2fbeb695fbaf8fc42704a06bf366b3b3c7fd8d21f2957"} Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.722535 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9sk5p" Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.881772 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m7pvk"] Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.890884 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m7pvk"] Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.905266 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9sk5p"] Sep 30 14:37:47 crc kubenswrapper[4799]: I0930 14:37:47.912824 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9sk5p"] Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.413231 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.413591 4799 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.413858 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g9g7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(ae8631cb-f4a0-4d95-afb4-64129fc0475a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.415578 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="ae8631cb-f4a0-4d95-afb4-64129fc0475a" Sep 30 14:37:48 crc kubenswrapper[4799]: I0930 14:37:48.531449 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df4d1b72-8dc1-4601-9766-1544375fb67e" path="/var/lib/kubelet/pods/df4d1b72-8dc1-4601-9766-1544375fb67e/volumes" Sep 30 14:37:48 crc kubenswrapper[4799]: I0930 14:37:48.532145 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1709326-a1b4-45ad-bb14-f3d0b2eb5124" path="/var/lib/kubelet/pods/e1709326-a1b4-45ad-bb14-f3d0b2eb5124/volumes" Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.645639 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = writing blob: storing blob to file \\\"/var/tmp/container_images_storage2740509720/2\\\": happened during read: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="8d8fe29c-5270-400a-9fc6-1024482dd54d" Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.678943 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="c146ba83-511e-4ff2-9a3e-6c83c4c43e9e" Sep 30 14:37:48 crc kubenswrapper[4799]: I0930 14:37:48.731136 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-r82xm" event={"ID":"9ee89b02-9e5e-4c62-ade3-c4a308feefde","Type":"ContainerStarted","Data":"e27df080aa26f959cb84288997c2495aca434229efb6fcbd39a9bebc49e6f18e"} Sep 30 14:37:48 crc kubenswrapper[4799]: I0930 14:37:48.732299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e","Type":"ContainerStarted","Data":"9b9d4293b019e420f8b67734172345621d7b78d8036d8c9b959d3562b70301a1"} Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.733443 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="c146ba83-511e-4ff2-9a3e-6c83c4c43e9e" Sep 30 14:37:48 crc kubenswrapper[4799]: I0930 14:37:48.734500 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b86436d8-96e2-44aa-9b4c-b9292a1ca129","Type":"ContainerStarted","Data":"c54e4c0284f48b211e22a3f7a912b188a8574b5b3681868401d3e202a500b907"} Sep 30 14:37:48 crc kubenswrapper[4799]: I0930 14:37:48.736740 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8d8fe29c-5270-400a-9fc6-1024482dd54d","Type":"ContainerStarted","Data":"4b37354a2f24323cdb2527423066867fc7aa320d0d55b86be5aea0ef65a4bcc1"} Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.738167 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="8d8fe29c-5270-400a-9fc6-1024482dd54d" Sep 30 14:37:48 crc kubenswrapper[4799]: I0930 14:37:48.742218 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"33a2e46f-eaef-4d54-8af4-ef388d3425dc","Type":"ContainerStarted","Data":"0958efab420e7f7d73d3511da528acfbe9bb062a33e441e75ad2b34f7ddd8dde"} Sep 30 14:37:48 crc kubenswrapper[4799]: E0930 14:37:48.744262 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="ae8631cb-f4a0-4d95-afb4-64129fc0475a" Sep 30 14:37:48 crc kubenswrapper[4799]: I0930 14:37:48.756228 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-r82xm" podStartSLOduration=-9223372004.098568 podStartE2EDuration="32.756208381s" podCreationTimestamp="2025-09-30 14:37:16 +0000 UTC" firstStartedPulling="2025-09-30 14:37:17.406786952 +0000 UTC m=+1059.490387379" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:37:48.753087469 +0000 UTC m=+1090.836687896" watchObservedRunningTime="2025-09-30 14:37:48.756208381 +0000 UTC m=+1090.839808808" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.207887 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rctd8"] Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.275878 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x866x"] Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.277773 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.290532 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x866x"] Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.290719 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.364722 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7tx\" (UniqueName: \"kubernetes.io/projected/3a3a0ca1-7c04-467f-8665-20bbac056d41-kube-api-access-dh7tx\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.364783 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.364846 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.364930 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-config\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.452470 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-l6mn9"] Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.466853 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.466914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7tx\" (UniqueName: \"kubernetes.io/projected/3a3a0ca1-7c04-467f-8665-20bbac056d41-kube-api-access-dh7tx\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.466968 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.467019 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-config\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.468048 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-config\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.468733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.469523 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.508719 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7tx\" (UniqueName: \"kubernetes.io/projected/3a3a0ca1-7c04-467f-8665-20bbac056d41-kube-api-access-dh7tx\") pod \"dnsmasq-dns-7f896c8c65-x866x\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.523855 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9jrdq"] Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.525646 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.540821 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.561134 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9jrdq"] Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.569107 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.569162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-config\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.569193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.569227 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.569256 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ks2h\" (UniqueName: \"kubernetes.io/projected/b653ad5b-ace8-4659-8f9c-af1bb901776f-kube-api-access-2ks2h\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.606484 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.677675 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.677769 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-config\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.677811 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.677851 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.677893 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ks2h\" (UniqueName: \"kubernetes.io/projected/b653ad5b-ace8-4659-8f9c-af1bb901776f-kube-api-access-2ks2h\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.679294 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-config\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.679325 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.679916 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.680121 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.713891 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ks2h\" (UniqueName: \"kubernetes.io/projected/b653ad5b-ace8-4659-8f9c-af1bb901776f-kube-api-access-2ks2h\") pod \"dnsmasq-dns-86db49b7ff-9jrdq\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.758816 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" event={"ID":"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df","Type":"ContainerDied","Data":"fdf8b194210af49b0144357c37680d05771954f1edb94c4a0201851b7e2d3ff0"} Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.758856 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdf8b194210af49b0144357c37680d05771954f1edb94c4a0201851b7e2d3ff0" Sep 30 14:37:49 crc kubenswrapper[4799]: E0930 14:37:49.760987 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="8d8fe29c-5270-400a-9fc6-1024482dd54d" Sep 30 14:37:49 crc kubenswrapper[4799]: E0930 14:37:49.761476 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="c146ba83-511e-4ff2-9a3e-6c83c4c43e9e" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.847093 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.872445 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.895364 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-dns-svc\") pod \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.895505 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrqr4\" (UniqueName: \"kubernetes.io/projected/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-kube-api-access-rrqr4\") pod \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.895635 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-config\") pod \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\" (UID: \"dde499c4-cff7-41b0-80b2-2a9ee3f7b0df\") " Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.899728 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-config" (OuterVolumeSpecName: "config") pod "dde499c4-cff7-41b0-80b2-2a9ee3f7b0df" (UID: "dde499c4-cff7-41b0-80b2-2a9ee3f7b0df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.900091 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dde499c4-cff7-41b0-80b2-2a9ee3f7b0df" (UID: "dde499c4-cff7-41b0-80b2-2a9ee3f7b0df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.905071 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-kube-api-access-rrqr4" (OuterVolumeSpecName: "kube-api-access-rrqr4") pod "dde499c4-cff7-41b0-80b2-2a9ee3f7b0df" (UID: "dde499c4-cff7-41b0-80b2-2a9ee3f7b0df"). InnerVolumeSpecName "kube-api-access-rrqr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.997904 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.997945 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:49 crc kubenswrapper[4799]: I0930 14:37:49.997958 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrqr4\" (UniqueName: \"kubernetes.io/projected/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df-kube-api-access-rrqr4\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.058403 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.202385 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnv4m\" (UniqueName: \"kubernetes.io/projected/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-kube-api-access-dnv4m\") pod \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.202807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-config\") pod \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.202964 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-dns-svc\") pod \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\" (UID: \"4adfc365-32d8-44a1-9eca-11eabcdc1ab6\") " Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.203522 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-config" (OuterVolumeSpecName: "config") pod "4adfc365-32d8-44a1-9eca-11eabcdc1ab6" (UID: "4adfc365-32d8-44a1-9eca-11eabcdc1ab6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.203805 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4adfc365-32d8-44a1-9eca-11eabcdc1ab6" (UID: "4adfc365-32d8-44a1-9eca-11eabcdc1ab6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.209954 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-kube-api-access-dnv4m" (OuterVolumeSpecName: "kube-api-access-dnv4m") pod "4adfc365-32d8-44a1-9eca-11eabcdc1ab6" (UID: "4adfc365-32d8-44a1-9eca-11eabcdc1ab6"). InnerVolumeSpecName "kube-api-access-dnv4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.296102 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x866x"] Sep 30 14:37:50 crc kubenswrapper[4799]: W0930 14:37:50.297510 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a3a0ca1_7c04_467f_8665_20bbac056d41.slice/crio-7b74362d4ddafa38bcc3bc564f8b6226d00b125dbba33d0cbc078675b9071f2b WatchSource:0}: Error finding container 7b74362d4ddafa38bcc3bc564f8b6226d00b125dbba33d0cbc078675b9071f2b: Status 404 returned error can't find the container with id 7b74362d4ddafa38bcc3bc564f8b6226d00b125dbba33d0cbc078675b9071f2b Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.304393 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.304437 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.304450 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnv4m\" (UniqueName: \"kubernetes.io/projected/4adfc365-32d8-44a1-9eca-11eabcdc1ab6-kube-api-access-dnv4m\") on node \"crc\" DevicePath \"\"" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.422703 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9jrdq"] Sep 30 14:37:50 crc kubenswrapper[4799]: W0930 14:37:50.423958 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb653ad5b_ace8_4659_8f9c_af1bb901776f.slice/crio-2db3943b54d96110da96e33ff95aa78e8fe0ee950c81a42257e92745f7cded6f WatchSource:0}: Error finding container 2db3943b54d96110da96e33ff95aa78e8fe0ee950c81a42257e92745f7cded6f: Status 404 returned error can't find the container with id 2db3943b54d96110da96e33ff95aa78e8fe0ee950c81a42257e92745f7cded6f Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.774932 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" event={"ID":"3a3a0ca1-7c04-467f-8665-20bbac056d41","Type":"ContainerStarted","Data":"7b74362d4ddafa38bcc3bc564f8b6226d00b125dbba33d0cbc078675b9071f2b"} Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.776610 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.776658 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-l6mn9" event={"ID":"4adfc365-32d8-44a1-9eca-11eabcdc1ab6","Type":"ContainerDied","Data":"aefffeddb19eb23a595487b08521d8cb65060aa71d463ef6d1cc12eccff8ace9"} Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.782398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" event={"ID":"b653ad5b-ace8-4659-8f9c-af1bb901776f","Type":"ContainerStarted","Data":"2db3943b54d96110da96e33ff95aa78e8fe0ee950c81a42257e92745f7cded6f"} Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.782448 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-rctd8" Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.934932 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-l6mn9"] Sep 30 14:37:50 crc kubenswrapper[4799]: I0930 14:37:50.945099 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-l6mn9"] Sep 30 14:37:51 crc kubenswrapper[4799]: I0930 14:37:51.011254 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rctd8"] Sep 30 14:37:51 crc kubenswrapper[4799]: I0930 14:37:51.019381 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rctd8"] Sep 30 14:37:52 crc kubenswrapper[4799]: I0930 14:37:52.514694 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4adfc365-32d8-44a1-9eca-11eabcdc1ab6" path="/var/lib/kubelet/pods/4adfc365-32d8-44a1-9eca-11eabcdc1ab6/volumes" Sep 30 14:37:52 crc kubenswrapper[4799]: I0930 14:37:52.515680 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dde499c4-cff7-41b0-80b2-2a9ee3f7b0df" path="/var/lib/kubelet/pods/dde499c4-cff7-41b0-80b2-2a9ee3f7b0df/volumes" Sep 30 14:37:52 crc kubenswrapper[4799]: I0930 14:37:52.800272 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" event={"ID":"3a3a0ca1-7c04-467f-8665-20bbac056d41","Type":"ContainerStarted","Data":"d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b"} Sep 30 14:37:53 crc kubenswrapper[4799]: I0930 14:37:53.811268 4799 generic.go:334] "Generic (PLEG): container finished" podID="3a3a0ca1-7c04-467f-8665-20bbac056d41" containerID="d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b" exitCode=0 Sep 30 14:37:53 crc kubenswrapper[4799]: I0930 14:37:53.811354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" event={"ID":"3a3a0ca1-7c04-467f-8665-20bbac056d41","Type":"ContainerDied","Data":"d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b"} Sep 30 14:37:53 crc kubenswrapper[4799]: I0930 14:37:53.816829 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" event={"ID":"b653ad5b-ace8-4659-8f9c-af1bb901776f","Type":"ContainerStarted","Data":"3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09"} Sep 30 14:37:54 crc kubenswrapper[4799]: I0930 14:37:54.825547 4799 generic.go:334] "Generic (PLEG): container finished" podID="b653ad5b-ace8-4659-8f9c-af1bb901776f" containerID="3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09" exitCode=0 Sep 30 14:37:54 crc kubenswrapper[4799]: I0930 14:37:54.825726 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" event={"ID":"b653ad5b-ace8-4659-8f9c-af1bb901776f","Type":"ContainerDied","Data":"3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09"} Sep 30 14:37:55 crc kubenswrapper[4799]: I0930 14:37:55.836915 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" event={"ID":"3a3a0ca1-7c04-467f-8665-20bbac056d41","Type":"ContainerStarted","Data":"922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7"} Sep 30 14:37:55 crc kubenswrapper[4799]: I0930 14:37:55.887109 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" podStartSLOduration=6.432084461 podStartE2EDuration="6.887080093s" podCreationTimestamp="2025-09-30 14:37:49 +0000 UTC" firstStartedPulling="2025-09-30 14:37:50.300421033 +0000 UTC m=+1092.384021460" lastFinishedPulling="2025-09-30 14:37:50.755416665 +0000 UTC m=+1092.839017092" observedRunningTime="2025-09-30 14:37:55.878686138 +0000 UTC m=+1097.962286585" watchObservedRunningTime="2025-09-30 14:37:55.887080093 +0000 UTC m=+1097.970680520" Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.847885 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb82f13-0450-42a8-9c8a-0c174c534382","Type":"ContainerStarted","Data":"0817b5a4428f559089509977d4a81d1d3084c69e64b252f2427d59eae955efa7"} Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.850482 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef6dcf7e-70a0-4759-afef-53e70b800ba0","Type":"ContainerStarted","Data":"4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3"} Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.853248 4799 generic.go:334] "Generic (PLEG): container finished" podID="b86436d8-96e2-44aa-9b4c-b9292a1ca129" containerID="c54e4c0284f48b211e22a3f7a912b188a8574b5b3681868401d3e202a500b907" exitCode=0 Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.853282 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b86436d8-96e2-44aa-9b4c-b9292a1ca129","Type":"ContainerDied","Data":"c54e4c0284f48b211e22a3f7a912b188a8574b5b3681868401d3e202a500b907"} Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.856828 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" event={"ID":"b653ad5b-ace8-4659-8f9c-af1bb901776f","Type":"ContainerStarted","Data":"ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b"} Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.857244 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.859929 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e3bd7a13-1fa3-4901-aeb6-3b9011ae330b","Type":"ContainerStarted","Data":"3734f8a2d84e3a0b31105cb398755e280ea74289487f510e6d255471c307b832"} Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.860210 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.860384 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.931968 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.116962076 podStartE2EDuration="48.931937369s" podCreationTimestamp="2025-09-30 14:37:08 +0000 UTC" firstStartedPulling="2025-09-30 14:37:10.200807223 +0000 UTC m=+1052.284407650" lastFinishedPulling="2025-09-30 14:37:56.015782506 +0000 UTC m=+1098.099382943" observedRunningTime="2025-09-30 14:37:56.925307155 +0000 UTC m=+1099.008907632" watchObservedRunningTime="2025-09-30 14:37:56.931937369 +0000 UTC m=+1099.015537796" Sep 30 14:37:56 crc kubenswrapper[4799]: I0930 14:37:56.979198 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" podStartSLOduration=7.527878086 podStartE2EDuration="7.979168349s" podCreationTimestamp="2025-09-30 14:37:49 +0000 UTC" firstStartedPulling="2025-09-30 14:37:50.426466478 +0000 UTC m=+1092.510066915" lastFinishedPulling="2025-09-30 14:37:50.877756751 +0000 UTC m=+1092.961357178" observedRunningTime="2025-09-30 14:37:56.973596906 +0000 UTC m=+1099.057197353" watchObservedRunningTime="2025-09-30 14:37:56.979168349 +0000 UTC m=+1099.062768776" Sep 30 14:37:57 crc kubenswrapper[4799]: I0930 14:37:57.870581 4799 generic.go:334] "Generic (PLEG): container finished" podID="33a2e46f-eaef-4d54-8af4-ef388d3425dc" containerID="0958efab420e7f7d73d3511da528acfbe9bb062a33e441e75ad2b34f7ddd8dde" exitCode=0 Sep 30 14:37:57 crc kubenswrapper[4799]: I0930 14:37:57.870676 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"33a2e46f-eaef-4d54-8af4-ef388d3425dc","Type":"ContainerDied","Data":"0958efab420e7f7d73d3511da528acfbe9bb062a33e441e75ad2b34f7ddd8dde"} Sep 30 14:37:57 crc kubenswrapper[4799]: I0930 14:37:57.877322 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b86436d8-96e2-44aa-9b4c-b9292a1ca129","Type":"ContainerStarted","Data":"6b6d432f924fb78da101fe1d28aceb2a68a2fc511713d876b0f8dd9e85d05040"} Sep 30 14:37:57 crc kubenswrapper[4799]: I0930 14:37:57.921139 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.3449175 podStartE2EDuration="51.921117626s" podCreationTimestamp="2025-09-30 14:37:06 +0000 UTC" firstStartedPulling="2025-09-30 14:37:09.902217474 +0000 UTC m=+1051.985817911" lastFinishedPulling="2025-09-30 14:37:48.47841761 +0000 UTC m=+1090.562018037" observedRunningTime="2025-09-30 14:37:57.918993184 +0000 UTC m=+1100.002593611" watchObservedRunningTime="2025-09-30 14:37:57.921117626 +0000 UTC m=+1100.004718053" Sep 30 14:37:58 crc kubenswrapper[4799]: I0930 14:37:58.061003 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Sep 30 14:37:58 crc kubenswrapper[4799]: I0930 14:37:58.061194 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Sep 30 14:37:59 crc kubenswrapper[4799]: I0930 14:37:59.649423 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:37:59 crc kubenswrapper[4799]: I0930 14:37:59.649811 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:38:01 crc kubenswrapper[4799]: I0930 14:38:01.911553 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v2rs4" event={"ID":"17bda665-adad-40d3-9465-46013020af4e","Type":"ContainerStarted","Data":"00efa256d3521cc65b745ac90ba23e7c2790b3ed2b42563ba0bd62985a3b9d6e"} Sep 30 14:38:01 crc kubenswrapper[4799]: I0930 14:38:01.914321 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"33a2e46f-eaef-4d54-8af4-ef388d3425dc","Type":"ContainerStarted","Data":"3da66a34cf6d6d82b1d3f79a74a2c2508f663b9a34ad41c62d8658ca0e66fd69"} Sep 30 14:38:02 crc kubenswrapper[4799]: I0930 14:38:02.926987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v2rs4" event={"ID":"17bda665-adad-40d3-9465-46013020af4e","Type":"ContainerDied","Data":"00efa256d3521cc65b745ac90ba23e7c2790b3ed2b42563ba0bd62985a3b9d6e"} Sep 30 14:38:02 crc kubenswrapper[4799]: I0930 14:38:02.926895 4799 generic.go:334] "Generic (PLEG): container finished" podID="17bda665-adad-40d3-9465-46013020af4e" containerID="00efa256d3521cc65b745ac90ba23e7c2790b3ed2b42563ba0bd62985a3b9d6e" exitCode=0 Sep 30 14:38:02 crc kubenswrapper[4799]: I0930 14:38:02.931867 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2kvml" event={"ID":"cbefa245-5ad5-4303-b166-59574e8a0c65","Type":"ContainerStarted","Data":"184cfe428e4a68f4f71f2f993b4c084f2026d86e1855040f4848e3e7737245db"} Sep 30 14:38:02 crc kubenswrapper[4799]: I0930 14:38:02.932339 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-2kvml" Sep 30 14:38:02 crc kubenswrapper[4799]: I0930 14:38:02.936670 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c146ba83-511e-4ff2-9a3e-6c83c4c43e9e","Type":"ContainerStarted","Data":"3d9aef676fdd9e506f05d29082fc11a1de3805b1a1d5aa95c9c14823e8d9cb21"} Sep 30 14:38:02 crc kubenswrapper[4799]: I0930 14:38:02.954669 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.224473348 podStartE2EDuration="56.954625015s" podCreationTimestamp="2025-09-30 14:37:06 +0000 UTC" firstStartedPulling="2025-09-30 14:37:09.691888855 +0000 UTC m=+1051.775489282" lastFinishedPulling="2025-09-30 14:37:48.422040522 +0000 UTC m=+1090.505640949" observedRunningTime="2025-09-30 14:38:01.976153091 +0000 UTC m=+1104.059753538" watchObservedRunningTime="2025-09-30 14:38:02.954625015 +0000 UTC m=+1105.038225442" Sep 30 14:38:02 crc kubenswrapper[4799]: I0930 14:38:02.989631 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-2kvml" podStartSLOduration=2.628884313 podStartE2EDuration="50.989600458s" podCreationTimestamp="2025-09-30 14:37:12 +0000 UTC" firstStartedPulling="2025-09-30 14:37:13.58924288 +0000 UTC m=+1055.672843307" lastFinishedPulling="2025-09-30 14:38:01.949959025 +0000 UTC m=+1104.033559452" observedRunningTime="2025-09-30 14:38:02.977949497 +0000 UTC m=+1105.061549944" watchObservedRunningTime="2025-09-30 14:38:02.989600458 +0000 UTC m=+1105.073200885" Sep 30 14:38:03 crc kubenswrapper[4799]: I0930 14:38:03.012045 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.577845633 podStartE2EDuration="50.012016513s" podCreationTimestamp="2025-09-30 14:37:13 +0000 UTC" firstStartedPulling="2025-09-30 14:37:16.885470942 +0000 UTC m=+1058.969071369" lastFinishedPulling="2025-09-30 14:38:02.319641822 +0000 UTC m=+1104.403242249" observedRunningTime="2025-09-30 14:38:03.007152201 +0000 UTC m=+1105.090752638" watchObservedRunningTime="2025-09-30 14:38:03.012016513 +0000 UTC m=+1105.095616940" Sep 30 14:38:03 crc kubenswrapper[4799]: I0930 14:38:03.752074 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Sep 30 14:38:03 crc kubenswrapper[4799]: I0930 14:38:03.947215 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8d8fe29c-5270-400a-9fc6-1024482dd54d","Type":"ContainerStarted","Data":"5659b02d27fc05448b11f63c520c0f5a15bcabe0e8bf1a53e6b066ac268a2222"} Sep 30 14:38:03 crc kubenswrapper[4799]: I0930 14:38:03.950574 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v2rs4" event={"ID":"17bda665-adad-40d3-9465-46013020af4e","Type":"ContainerStarted","Data":"8ef9cdaa447a4034123bfc22741fe4cdbbb9fd76bb040fe461cd11d46d6f1eb1"} Sep 30 14:38:03 crc kubenswrapper[4799]: I0930 14:38:03.950832 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v2rs4" event={"ID":"17bda665-adad-40d3-9465-46013020af4e","Type":"ContainerStarted","Data":"5aae479dcfa653d70f2c42bbfdbc9094953d1db538cebf3759faf4a388505969"} Sep 30 14:38:03 crc kubenswrapper[4799]: I0930 14:38:03.951052 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:38:03 crc kubenswrapper[4799]: I0930 14:38:03.951168 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:38:03 crc kubenswrapper[4799]: I0930 14:38:03.991799 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.157088799 podStartE2EDuration="47.991762115s" podCreationTimestamp="2025-09-30 14:37:16 +0000 UTC" firstStartedPulling="2025-09-30 14:37:23.686202343 +0000 UTC m=+1065.769802770" lastFinishedPulling="2025-09-30 14:38:03.520875659 +0000 UTC m=+1105.604476086" observedRunningTime="2025-09-30 14:38:03.986943824 +0000 UTC m=+1106.070544261" watchObservedRunningTime="2025-09-30 14:38:03.991762115 +0000 UTC m=+1106.075362542" Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.033821 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-v2rs4" podStartSLOduration=6.851284459 podStartE2EDuration="52.033783123s" podCreationTimestamp="2025-09-30 14:37:12 +0000 UTC" firstStartedPulling="2025-09-30 14:37:16.319862756 +0000 UTC m=+1058.403463183" lastFinishedPulling="2025-09-30 14:38:01.50236141 +0000 UTC m=+1103.585961847" observedRunningTime="2025-09-30 14:38:04.026426748 +0000 UTC m=+1106.110027195" watchObservedRunningTime="2025-09-30 14:38:04.033783123 +0000 UTC m=+1106.117383560" Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.310126 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.413866 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.516630 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.609079 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.873876 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.933290 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x866x"] Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.960179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ae8631cb-f4a0-4d95-afb4-64129fc0475a","Type":"ContainerStarted","Data":"0bec0a419315cfe668fcfe77877d66678738b2af950136f5bd931bf40eed6749"} Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.960400 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" podUID="3a3a0ca1-7c04-467f-8665-20bbac056d41" containerName="dnsmasq-dns" containerID="cri-o://922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7" gracePeriod=10 Sep 30 14:38:04 crc kubenswrapper[4799]: I0930 14:38:04.994113 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.890239614 podStartE2EDuration="54.994086596s" podCreationTimestamp="2025-09-30 14:37:10 +0000 UTC" firstStartedPulling="2025-09-30 14:37:11.939248594 +0000 UTC m=+1054.022849021" lastFinishedPulling="2025-09-30 14:38:04.043095576 +0000 UTC m=+1106.126696003" observedRunningTime="2025-09-30 14:38:04.990625895 +0000 UTC m=+1107.074226342" watchObservedRunningTime="2025-09-30 14:38:04.994086596 +0000 UTC m=+1107.077687033" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.436710 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.514252 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.567576 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.581835 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh7tx\" (UniqueName: \"kubernetes.io/projected/3a3a0ca1-7c04-467f-8665-20bbac056d41-kube-api-access-dh7tx\") pod \"3a3a0ca1-7c04-467f-8665-20bbac056d41\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.581918 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-ovsdbserver-sb\") pod \"3a3a0ca1-7c04-467f-8665-20bbac056d41\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.582076 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-config\") pod \"3a3a0ca1-7c04-467f-8665-20bbac056d41\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.582176 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-dns-svc\") pod \"3a3a0ca1-7c04-467f-8665-20bbac056d41\" (UID: \"3a3a0ca1-7c04-467f-8665-20bbac056d41\") " Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.596006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3a0ca1-7c04-467f-8665-20bbac056d41-kube-api-access-dh7tx" (OuterVolumeSpecName: "kube-api-access-dh7tx") pod "3a3a0ca1-7c04-467f-8665-20bbac056d41" (UID: "3a3a0ca1-7c04-467f-8665-20bbac056d41"). InnerVolumeSpecName "kube-api-access-dh7tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.639922 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3a3a0ca1-7c04-467f-8665-20bbac056d41" (UID: "3a3a0ca1-7c04-467f-8665-20bbac056d41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.669514 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.670777 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-config" (OuterVolumeSpecName: "config") pod "3a3a0ca1-7c04-467f-8665-20bbac056d41" (UID: "3a3a0ca1-7c04-467f-8665-20bbac056d41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.671186 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3a3a0ca1-7c04-467f-8665-20bbac056d41" (UID: "3a3a0ca1-7c04-467f-8665-20bbac056d41"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.684441 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh7tx\" (UniqueName: \"kubernetes.io/projected/3a3a0ca1-7c04-467f-8665-20bbac056d41-kube-api-access-dh7tx\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.684480 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.684490 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.684499 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a3a0ca1-7c04-467f-8665-20bbac056d41-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.970384 4799 generic.go:334] "Generic (PLEG): container finished" podID="3a3a0ca1-7c04-467f-8665-20bbac056d41" containerID="922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7" exitCode=0 Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.971936 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.972823 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" event={"ID":"3a3a0ca1-7c04-467f-8665-20bbac056d41","Type":"ContainerDied","Data":"922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7"} Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.972956 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-x866x" event={"ID":"3a3a0ca1-7c04-467f-8665-20bbac056d41","Type":"ContainerDied","Data":"7b74362d4ddafa38bcc3bc564f8b6226d00b125dbba33d0cbc078675b9071f2b"} Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.972990 4799 scope.go:117] "RemoveContainer" containerID="922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7" Sep 30 14:38:05 crc kubenswrapper[4799]: I0930 14:38:05.998798 4799 scope.go:117] "RemoveContainer" containerID="d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b" Sep 30 14:38:06 crc kubenswrapper[4799]: I0930 14:38:06.017422 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x866x"] Sep 30 14:38:06 crc kubenswrapper[4799]: I0930 14:38:06.023218 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-x866x"] Sep 30 14:38:06 crc kubenswrapper[4799]: I0930 14:38:06.026858 4799 scope.go:117] "RemoveContainer" containerID="922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7" Sep 30 14:38:06 crc kubenswrapper[4799]: E0930 14:38:06.027938 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7\": container with ID starting with 922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7 not found: ID does not exist" containerID="922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7" Sep 30 14:38:06 crc kubenswrapper[4799]: I0930 14:38:06.027978 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7"} err="failed to get container status \"922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7\": rpc error: code = NotFound desc = could not find container \"922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7\": container with ID starting with 922644a8070b75843736319375edfdcbcd84582689e5038b089efd6207f40ad7 not found: ID does not exist" Sep 30 14:38:06 crc kubenswrapper[4799]: I0930 14:38:06.028009 4799 scope.go:117] "RemoveContainer" containerID="d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b" Sep 30 14:38:06 crc kubenswrapper[4799]: E0930 14:38:06.028474 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b\": container with ID starting with d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b not found: ID does not exist" containerID="d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b" Sep 30 14:38:06 crc kubenswrapper[4799]: I0930 14:38:06.028500 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b"} err="failed to get container status \"d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b\": rpc error: code = NotFound desc = could not find container \"d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b\": container with ID starting with d1ae5f39befd0f3c8610c025a7eb01a7d3d99bd3b8e029bd01f84e1c85734f5b not found: ID does not exist" Sep 30 14:38:06 crc kubenswrapper[4799]: I0930 14:38:06.515909 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a3a0ca1-7c04-467f-8665-20bbac056d41" path="/var/lib/kubelet/pods/3a3a0ca1-7c04-467f-8665-20bbac056d41/volumes" Sep 30 14:38:07 crc kubenswrapper[4799]: I0930 14:38:07.670331 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.144161 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-4tjtd"] Sep 30 14:38:08 crc kubenswrapper[4799]: E0930 14:38:08.145150 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3a0ca1-7c04-467f-8665-20bbac056d41" containerName="dnsmasq-dns" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.145275 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3a0ca1-7c04-467f-8665-20bbac056d41" containerName="dnsmasq-dns" Sep 30 14:38:08 crc kubenswrapper[4799]: E0930 14:38:08.145356 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3a0ca1-7c04-467f-8665-20bbac056d41" containerName="init" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.145425 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3a0ca1-7c04-467f-8665-20bbac056d41" containerName="init" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.146577 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a3a0ca1-7c04-467f-8665-20bbac056d41" containerName="dnsmasq-dns" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.147352 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.147375 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.147885 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4tjtd" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.165390 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4tjtd"] Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.210316 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.227061 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jgqg\" (UniqueName: \"kubernetes.io/projected/c6f59e8e-8173-468f-81c8-9f86067de6ea-kube-api-access-2jgqg\") pod \"keystone-db-create-4tjtd\" (UID: \"c6f59e8e-8173-468f-81c8-9f86067de6ea\") " pod="openstack/keystone-db-create-4tjtd" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.328787 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jgqg\" (UniqueName: \"kubernetes.io/projected/c6f59e8e-8173-468f-81c8-9f86067de6ea-kube-api-access-2jgqg\") pod \"keystone-db-create-4tjtd\" (UID: \"c6f59e8e-8173-468f-81c8-9f86067de6ea\") " pod="openstack/keystone-db-create-4tjtd" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.350387 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jgqg\" (UniqueName: \"kubernetes.io/projected/c6f59e8e-8173-468f-81c8-9f86067de6ea-kube-api-access-2jgqg\") pod \"keystone-db-create-4tjtd\" (UID: \"c6f59e8e-8173-468f-81c8-9f86067de6ea\") " pod="openstack/keystone-db-create-4tjtd" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.471138 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4tjtd" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.722719 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.790367 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.792528 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-rqtlt"] Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.793824 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rqtlt" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.801640 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rqtlt"] Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.942455 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vj8h\" (UniqueName: \"kubernetes.io/projected/ccab43ff-518e-4785-9234-477293d0da8b-kube-api-access-7vj8h\") pod \"placement-db-create-rqtlt\" (UID: \"ccab43ff-518e-4785-9234-477293d0da8b\") " pod="openstack/placement-db-create-rqtlt" Sep 30 14:38:08 crc kubenswrapper[4799]: I0930 14:38:08.945910 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4tjtd"] Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.012356 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4tjtd" event={"ID":"c6f59e8e-8173-468f-81c8-9f86067de6ea","Type":"ContainerStarted","Data":"1d0a7b1ffb63375d4b0bba3aa3f91c07f740209df606b1b1d86f8967934ff8b3"} Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.044139 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vj8h\" (UniqueName: \"kubernetes.io/projected/ccab43ff-518e-4785-9234-477293d0da8b-kube-api-access-7vj8h\") pod \"placement-db-create-rqtlt\" (UID: \"ccab43ff-518e-4785-9234-477293d0da8b\") " pod="openstack/placement-db-create-rqtlt" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.071419 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vj8h\" (UniqueName: \"kubernetes.io/projected/ccab43ff-518e-4785-9234-477293d0da8b-kube-api-access-7vj8h\") pod \"placement-db-create-rqtlt\" (UID: \"ccab43ff-518e-4785-9234-477293d0da8b\") " pod="openstack/placement-db-create-rqtlt" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.107782 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.127233 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rqtlt" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.566580 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.681214 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rqtlt"] Sep 30 14:38:09 crc kubenswrapper[4799]: W0930 14:38:09.686068 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccab43ff_518e_4785_9234_477293d0da8b.slice/crio-7a88ceb1ebdbc7f004bace4c0fa9798cbb3406bdfe354b739622092a1feef36e WatchSource:0}: Error finding container 7a88ceb1ebdbc7f004bace4c0fa9798cbb3406bdfe354b739622092a1feef36e: Status 404 returned error can't find the container with id 7a88ceb1ebdbc7f004bace4c0fa9798cbb3406bdfe354b739622092a1feef36e Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.893029 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.894943 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.897373 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.898516 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-xxxfs" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.898886 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.899143 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.925269 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.960453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/583080ea-2fdf-461a-8b59-f04c48054eb9-config\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.960532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.960591 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxv58\" (UniqueName: \"kubernetes.io/projected/583080ea-2fdf-461a-8b59-f04c48054eb9-kube-api-access-kxv58\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.960621 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/583080ea-2fdf-461a-8b59-f04c48054eb9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.960701 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/583080ea-2fdf-461a-8b59-f04c48054eb9-scripts\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.960726 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:09 crc kubenswrapper[4799]: I0930 14:38:09.960757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.030521 4799 generic.go:334] "Generic (PLEG): container finished" podID="c6f59e8e-8173-468f-81c8-9f86067de6ea" containerID="b9c74512e48d404333cddb32ef0922c45770ad28d46a1af0a7d44a46a4307d38" exitCode=0 Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.031175 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4tjtd" event={"ID":"c6f59e8e-8173-468f-81c8-9f86067de6ea","Type":"ContainerDied","Data":"b9c74512e48d404333cddb32ef0922c45770ad28d46a1af0a7d44a46a4307d38"} Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.042234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rqtlt" event={"ID":"ccab43ff-518e-4785-9234-477293d0da8b","Type":"ContainerStarted","Data":"a17141d687adedf845311a4a394b436b1c880cf58ca3b23dc94d4706d2346cbf"} Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.042336 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rqtlt" event={"ID":"ccab43ff-518e-4785-9234-477293d0da8b","Type":"ContainerStarted","Data":"7a88ceb1ebdbc7f004bace4c0fa9798cbb3406bdfe354b739622092a1feef36e"} Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.067082 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/583080ea-2fdf-461a-8b59-f04c48054eb9-config\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.067149 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.067179 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxv58\" (UniqueName: \"kubernetes.io/projected/583080ea-2fdf-461a-8b59-f04c48054eb9-kube-api-access-kxv58\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.067198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/583080ea-2fdf-461a-8b59-f04c48054eb9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.067238 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/583080ea-2fdf-461a-8b59-f04c48054eb9-scripts\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.067253 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.067273 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.070555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/583080ea-2fdf-461a-8b59-f04c48054eb9-config\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.073414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/583080ea-2fdf-461a-8b59-f04c48054eb9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.074310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/583080ea-2fdf-461a-8b59-f04c48054eb9-scripts\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.081419 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.086236 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.090167 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxv58\" (UniqueName: \"kubernetes.io/projected/583080ea-2fdf-461a-8b59-f04c48054eb9-kube-api-access-kxv58\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.099842 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/583080ea-2fdf-461a-8b59-f04c48054eb9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"583080ea-2fdf-461a-8b59-f04c48054eb9\") " pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.215230 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.625159 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-rqtlt" podStartSLOduration=2.625127817 podStartE2EDuration="2.625127817s" podCreationTimestamp="2025-09-30 14:38:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:38:10.072557152 +0000 UTC m=+1112.156157599" watchObservedRunningTime="2025-09-30 14:38:10.625127817 +0000 UTC m=+1112.708728244" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.626686 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.963404 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 30 14:38:10 crc kubenswrapper[4799]: I0930 14:38:10.977511 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.074013 4799 generic.go:334] "Generic (PLEG): container finished" podID="ccab43ff-518e-4785-9234-477293d0da8b" containerID="a17141d687adedf845311a4a394b436b1c880cf58ca3b23dc94d4706d2346cbf" exitCode=0 Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.074214 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rqtlt" event={"ID":"ccab43ff-518e-4785-9234-477293d0da8b","Type":"ContainerDied","Data":"a17141d687adedf845311a4a394b436b1c880cf58ca3b23dc94d4706d2346cbf"} Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.085568 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"583080ea-2fdf-461a-8b59-f04c48054eb9","Type":"ContainerStarted","Data":"3d0c8c94bc6897601618bc6d959e66842512d4e660f3a3af8fc179faa7ea3fce"} Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.140959 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-kbqg4"] Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.142730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.182001 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kbqg4"] Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.199193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-config\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.199297 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvnxw\" (UniqueName: \"kubernetes.io/projected/86f6895b-e518-4d47-8b9e-ad1143dbc433-kube-api-access-pvnxw\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.199333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.199360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-dns-svc\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.211718 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.316531 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvnxw\" (UniqueName: \"kubernetes.io/projected/86f6895b-e518-4d47-8b9e-ad1143dbc433-kube-api-access-pvnxw\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.316587 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.316616 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-dns-svc\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.316736 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.316820 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-config\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.317770 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-config\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.318473 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-dns-svc\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.318812 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.317770 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.378057 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvnxw\" (UniqueName: \"kubernetes.io/projected/86f6895b-e518-4d47-8b9e-ad1143dbc433-kube-api-access-pvnxw\") pod \"dnsmasq-dns-698758b865-kbqg4\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:11 crc kubenswrapper[4799]: I0930 14:38:11.504749 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.011333 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4tjtd" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.117036 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4tjtd" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.117484 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4tjtd" event={"ID":"c6f59e8e-8173-468f-81c8-9f86067de6ea","Type":"ContainerDied","Data":"1d0a7b1ffb63375d4b0bba3aa3f91c07f740209df606b1b1d86f8967934ff8b3"} Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.117526 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d0a7b1ffb63375d4b0bba3aa3f91c07f740209df606b1b1d86f8967934ff8b3" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.144924 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jgqg\" (UniqueName: \"kubernetes.io/projected/c6f59e8e-8173-468f-81c8-9f86067de6ea-kube-api-access-2jgqg\") pod \"c6f59e8e-8173-468f-81c8-9f86067de6ea\" (UID: \"c6f59e8e-8173-468f-81c8-9f86067de6ea\") " Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.159738 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6f59e8e-8173-468f-81c8-9f86067de6ea-kube-api-access-2jgqg" (OuterVolumeSpecName: "kube-api-access-2jgqg") pod "c6f59e8e-8173-468f-81c8-9f86067de6ea" (UID: "c6f59e8e-8173-468f-81c8-9f86067de6ea"). InnerVolumeSpecName "kube-api-access-2jgqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.248415 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jgqg\" (UniqueName: \"kubernetes.io/projected/c6f59e8e-8173-468f-81c8-9f86067de6ea-kube-api-access-2jgqg\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.347587 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Sep 30 14:38:12 crc kubenswrapper[4799]: E0930 14:38:12.348107 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6f59e8e-8173-468f-81c8-9f86067de6ea" containerName="mariadb-database-create" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.348126 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6f59e8e-8173-468f-81c8-9f86067de6ea" containerName="mariadb-database-create" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.348358 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6f59e8e-8173-468f-81c8-9f86067de6ea" containerName="mariadb-database-create" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.354585 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.358501 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.367052 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-9k5b6" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.367361 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.367547 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.367621 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.438233 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kbqg4"] Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.451939 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.452269 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.453169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1500572d-64d4-483a-a525-bc0530fe4cbb-lock\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.453302 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8chf4\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-kube-api-access-8chf4\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.454362 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1500572d-64d4-483a-a525-bc0530fe4cbb-cache\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.556052 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1500572d-64d4-483a-a525-bc0530fe4cbb-cache\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.556143 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.556200 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.556261 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1500572d-64d4-483a-a525-bc0530fe4cbb-lock\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.556292 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8chf4\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-kube-api-access-8chf4\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: E0930 14:38:12.559771 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:38:12 crc kubenswrapper[4799]: E0930 14:38:12.559800 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:38:12 crc kubenswrapper[4799]: E0930 14:38:12.560058 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift podName:1500572d-64d4-483a-a525-bc0530fe4cbb nodeName:}" failed. No retries permitted until 2025-09-30 14:38:13.060034161 +0000 UTC m=+1115.143634588 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift") pod "swift-storage-0" (UID: "1500572d-64d4-483a-a525-bc0530fe4cbb") : configmap "swift-ring-files" not found Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.559790 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.560826 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1500572d-64d4-483a-a525-bc0530fe4cbb-lock\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.563007 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1500572d-64d4-483a-a525-bc0530fe4cbb-cache\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.580757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8chf4\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-kube-api-access-8chf4\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.607609 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.751822 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-26sfw"] Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.753281 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.759926 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.759947 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.760049 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.808833 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-26sfw"] Sep 30 14:38:12 crc kubenswrapper[4799]: E0930 14:38:12.815918 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-cf9gf ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-26sfw" podUID="45ce7073-b801-44b8-b4cb-d8677c484640" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.828537 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-26sfw"] Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.887510 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-mbljj"] Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.896970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-swiftconf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.897564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-dispersionconf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.898148 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/45ce7073-b801-44b8-b4cb-d8677c484640-etc-swift\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.898349 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-combined-ca-bundle\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.898413 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-scripts\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.900272 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf9gf\" (UniqueName: \"kubernetes.io/projected/45ce7073-b801-44b8-b4cb-d8677c484640-kube-api-access-cf9gf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.900378 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-ring-data-devices\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.904277 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:12 crc kubenswrapper[4799]: I0930 14:38:12.913787 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mbljj"] Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-ring-data-devices\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002422 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-dispersionconf\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002457 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-ring-data-devices\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-swiftconf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002524 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-scripts\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002551 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-dispersionconf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002573 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/45ce7073-b801-44b8-b4cb-d8677c484640-etc-swift\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002799 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmtvm\" (UniqueName: \"kubernetes.io/projected/ad219d29-0576-4b77-b5b9-fc2c980b7093-kube-api-access-qmtvm\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.002940 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-combined-ca-bundle\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.003079 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-scripts\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.003183 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-ring-data-devices\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.003192 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-swiftconf\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.003113 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/45ce7073-b801-44b8-b4cb-d8677c484640-etc-swift\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.003453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-combined-ca-bundle\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.003522 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad219d29-0576-4b77-b5b9-fc2c980b7093-etc-swift\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.003608 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf9gf\" (UniqueName: \"kubernetes.io/projected/45ce7073-b801-44b8-b4cb-d8677c484640-kube-api-access-cf9gf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.003949 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-scripts\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.008406 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-combined-ca-bundle\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.009331 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-swiftconf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.012517 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-dispersionconf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.023611 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf9gf\" (UniqueName: \"kubernetes.io/projected/45ce7073-b801-44b8-b4cb-d8677c484640-kube-api-access-cf9gf\") pod \"swift-ring-rebalance-26sfw\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.100141 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rqtlt" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.109447 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vj8h\" (UniqueName: \"kubernetes.io/projected/ccab43ff-518e-4785-9234-477293d0da8b-kube-api-access-7vj8h\") pod \"ccab43ff-518e-4785-9234-477293d0da8b\" (UID: \"ccab43ff-518e-4785-9234-477293d0da8b\") " Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.109882 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.109941 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-swiftconf\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.109980 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-combined-ca-bundle\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.110010 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad219d29-0576-4b77-b5b9-fc2c980b7093-etc-swift\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.110067 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-dispersionconf\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.110088 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-ring-data-devices\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.110122 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-scripts\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.110177 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmtvm\" (UniqueName: \"kubernetes.io/projected/ad219d29-0576-4b77-b5b9-fc2c980b7093-kube-api-access-qmtvm\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.111459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad219d29-0576-4b77-b5b9-fc2c980b7093-etc-swift\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: E0930 14:38:13.111677 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:38:13 crc kubenswrapper[4799]: E0930 14:38:13.111696 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:38:13 crc kubenswrapper[4799]: E0930 14:38:13.111742 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift podName:1500572d-64d4-483a-a525-bc0530fe4cbb nodeName:}" failed. No retries permitted until 2025-09-30 14:38:14.111724509 +0000 UTC m=+1116.195324936 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift") pod "swift-storage-0" (UID: "1500572d-64d4-483a-a525-bc0530fe4cbb") : configmap "swift-ring-files" not found Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.115053 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-ring-data-devices\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.115953 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-swiftconf\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.116941 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-scripts\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.120482 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-combined-ca-bundle\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.135293 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccab43ff-518e-4785-9234-477293d0da8b-kube-api-access-7vj8h" (OuterVolumeSpecName: "kube-api-access-7vj8h") pod "ccab43ff-518e-4785-9234-477293d0da8b" (UID: "ccab43ff-518e-4785-9234-477293d0da8b"). InnerVolumeSpecName "kube-api-access-7vj8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.151262 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-dispersionconf\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.151346 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmtvm\" (UniqueName: \"kubernetes.io/projected/ad219d29-0576-4b77-b5b9-fc2c980b7093-kube-api-access-qmtvm\") pod \"swift-ring-rebalance-mbljj\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.155291 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rqtlt" event={"ID":"ccab43ff-518e-4785-9234-477293d0da8b","Type":"ContainerDied","Data":"7a88ceb1ebdbc7f004bace4c0fa9798cbb3406bdfe354b739622092a1feef36e"} Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.155351 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a88ceb1ebdbc7f004bace4c0fa9798cbb3406bdfe354b739622092a1feef36e" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.155455 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rqtlt" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.160274 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.160462 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kbqg4" event={"ID":"86f6895b-e518-4d47-8b9e-ad1143dbc433","Type":"ContainerStarted","Data":"378a5510f7c553447365e6cc0f9f777870de06ed4a7498339fdaa58d321c039c"} Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.202791 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.211764 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-combined-ca-bundle\") pod \"45ce7073-b801-44b8-b4cb-d8677c484640\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.212149 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-swiftconf\") pod \"45ce7073-b801-44b8-b4cb-d8677c484640\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.212263 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-scripts\") pod \"45ce7073-b801-44b8-b4cb-d8677c484640\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.212306 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-ring-data-devices\") pod \"45ce7073-b801-44b8-b4cb-d8677c484640\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.212429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/45ce7073-b801-44b8-b4cb-d8677c484640-etc-swift\") pod \"45ce7073-b801-44b8-b4cb-d8677c484640\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.212584 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-dispersionconf\") pod \"45ce7073-b801-44b8-b4cb-d8677c484640\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.212687 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf9gf\" (UniqueName: \"kubernetes.io/projected/45ce7073-b801-44b8-b4cb-d8677c484640-kube-api-access-cf9gf\") pod \"45ce7073-b801-44b8-b4cb-d8677c484640\" (UID: \"45ce7073-b801-44b8-b4cb-d8677c484640\") " Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.213448 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vj8h\" (UniqueName: \"kubernetes.io/projected/ccab43ff-518e-4785-9234-477293d0da8b-kube-api-access-7vj8h\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.214288 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "45ce7073-b801-44b8-b4cb-d8677c484640" (UID: "45ce7073-b801-44b8-b4cb-d8677c484640"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.214760 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-scripts" (OuterVolumeSpecName: "scripts") pod "45ce7073-b801-44b8-b4cb-d8677c484640" (UID: "45ce7073-b801-44b8-b4cb-d8677c484640"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.215066 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45ce7073-b801-44b8-b4cb-d8677c484640-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "45ce7073-b801-44b8-b4cb-d8677c484640" (UID: "45ce7073-b801-44b8-b4cb-d8677c484640"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.219551 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45ce7073-b801-44b8-b4cb-d8677c484640-kube-api-access-cf9gf" (OuterVolumeSpecName: "kube-api-access-cf9gf") pod "45ce7073-b801-44b8-b4cb-d8677c484640" (UID: "45ce7073-b801-44b8-b4cb-d8677c484640"). InnerVolumeSpecName "kube-api-access-cf9gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.222992 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "45ce7073-b801-44b8-b4cb-d8677c484640" (UID: "45ce7073-b801-44b8-b4cb-d8677c484640"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.224928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "45ce7073-b801-44b8-b4cb-d8677c484640" (UID: "45ce7073-b801-44b8-b4cb-d8677c484640"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.224979 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45ce7073-b801-44b8-b4cb-d8677c484640" (UID: "45ce7073-b801-44b8-b4cb-d8677c484640"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.244187 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.315721 4799 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-dispersionconf\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.315781 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf9gf\" (UniqueName: \"kubernetes.io/projected/45ce7073-b801-44b8-b4cb-d8677c484640-kube-api-access-cf9gf\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.315795 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.315807 4799 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/45ce7073-b801-44b8-b4cb-d8677c484640-swiftconf\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.315816 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.315824 4799 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/45ce7073-b801-44b8-b4cb-d8677c484640-ring-data-devices\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.315833 4799 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/45ce7073-b801-44b8-b4cb-d8677c484640-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:13 crc kubenswrapper[4799]: I0930 14:38:13.815386 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mbljj"] Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.018386 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-kv6wf"] Sep 30 14:38:14 crc kubenswrapper[4799]: E0930 14:38:14.019062 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccab43ff-518e-4785-9234-477293d0da8b" containerName="mariadb-database-create" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.019133 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccab43ff-518e-4785-9234-477293d0da8b" containerName="mariadb-database-create" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.019388 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccab43ff-518e-4785-9234-477293d0da8b" containerName="mariadb-database-create" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.020105 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kv6wf" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.036834 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kv6wf"] Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.145579 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48plr\" (UniqueName: \"kubernetes.io/projected/4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b-kube-api-access-48plr\") pod \"glance-db-create-kv6wf\" (UID: \"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b\") " pod="openstack/glance-db-create-kv6wf" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.145748 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:14 crc kubenswrapper[4799]: E0930 14:38:14.145973 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:38:14 crc kubenswrapper[4799]: E0930 14:38:14.145992 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:38:14 crc kubenswrapper[4799]: E0930 14:38:14.146059 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift podName:1500572d-64d4-483a-a525-bc0530fe4cbb nodeName:}" failed. No retries permitted until 2025-09-30 14:38:16.146031997 +0000 UTC m=+1118.229632454 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift") pod "swift-storage-0" (UID: "1500572d-64d4-483a-a525-bc0530fe4cbb") : configmap "swift-ring-files" not found Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.173338 4799 generic.go:334] "Generic (PLEG): container finished" podID="86f6895b-e518-4d47-8b9e-ad1143dbc433" containerID="df84b6ea18e8be02b814045e5d16038b565e6b0ba8ea61cadfc5e62f05c13ace" exitCode=0 Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.173447 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kbqg4" event={"ID":"86f6895b-e518-4d47-8b9e-ad1143dbc433","Type":"ContainerDied","Data":"df84b6ea18e8be02b814045e5d16038b565e6b0ba8ea61cadfc5e62f05c13ace"} Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.176478 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mbljj" event={"ID":"ad219d29-0576-4b77-b5b9-fc2c980b7093","Type":"ContainerStarted","Data":"8c1f7b80c591f62727a45edeb472f2b4f4b10a6fd2e633bd8cd0c34d034e480d"} Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.182799 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-26sfw" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.182888 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"583080ea-2fdf-461a-8b59-f04c48054eb9","Type":"ContainerStarted","Data":"982950ee463e4e6a295b0b7f21bb4dc42ea29a0474cabd57ca64ff053d0f0633"} Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.182933 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"583080ea-2fdf-461a-8b59-f04c48054eb9","Type":"ContainerStarted","Data":"f6e1bbb4640e81a994932504b319ab0a3964fb0c6bcfd2ee9d6acdebefa2735c"} Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.239239 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.470438649 podStartE2EDuration="5.239207581s" podCreationTimestamp="2025-09-30 14:38:09 +0000 UTC" firstStartedPulling="2025-09-30 14:38:10.653605009 +0000 UTC m=+1112.737205436" lastFinishedPulling="2025-09-30 14:38:13.422373941 +0000 UTC m=+1115.505974368" observedRunningTime="2025-09-30 14:38:14.235494512 +0000 UTC m=+1116.319094939" watchObservedRunningTime="2025-09-30 14:38:14.239207581 +0000 UTC m=+1116.322808018" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.249096 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48plr\" (UniqueName: \"kubernetes.io/projected/4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b-kube-api-access-48plr\") pod \"glance-db-create-kv6wf\" (UID: \"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b\") " pod="openstack/glance-db-create-kv6wf" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.287838 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-26sfw"] Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.292499 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48plr\" (UniqueName: \"kubernetes.io/projected/4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b-kube-api-access-48plr\") pod \"glance-db-create-kv6wf\" (UID: \"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b\") " pod="openstack/glance-db-create-kv6wf" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.295572 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-26sfw"] Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.344019 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kv6wf" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.527118 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45ce7073-b801-44b8-b4cb-d8677c484640" path="/var/lib/kubelet/pods/45ce7073-b801-44b8-b4cb-d8677c484640/volumes" Sep 30 14:38:14 crc kubenswrapper[4799]: I0930 14:38:14.928669 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kv6wf"] Sep 30 14:38:15 crc kubenswrapper[4799]: I0930 14:38:15.193534 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kv6wf" event={"ID":"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b","Type":"ContainerStarted","Data":"34e9f5d1a8a89605dc771a739f141bfb3191b1f19adba7d51b71aa26eae4c0ab"} Sep 30 14:38:15 crc kubenswrapper[4799]: I0930 14:38:15.194026 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kv6wf" event={"ID":"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b","Type":"ContainerStarted","Data":"cf9464b3dde9304cbec22f746b905fb61962b3ad1936005da38b4d5c0089b5d9"} Sep 30 14:38:15 crc kubenswrapper[4799]: I0930 14:38:15.204759 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kbqg4" event={"ID":"86f6895b-e518-4d47-8b9e-ad1143dbc433","Type":"ContainerStarted","Data":"0230a08df166e7eec4cbc204bd89968f918187177b7c0d95489950ceb9f902f8"} Sep 30 14:38:15 crc kubenswrapper[4799]: I0930 14:38:15.207579 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Sep 30 14:38:15 crc kubenswrapper[4799]: I0930 14:38:15.207756 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:15 crc kubenswrapper[4799]: I0930 14:38:15.211715 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-kv6wf" podStartSLOduration=2.211360979 podStartE2EDuration="2.211360979s" podCreationTimestamp="2025-09-30 14:38:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:38:15.208710302 +0000 UTC m=+1117.292310739" watchObservedRunningTime="2025-09-30 14:38:15.211360979 +0000 UTC m=+1117.294961406" Sep 30 14:38:15 crc kubenswrapper[4799]: I0930 14:38:15.245180 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-kbqg4" podStartSLOduration=4.24488774 podStartE2EDuration="4.24488774s" podCreationTimestamp="2025-09-30 14:38:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:38:15.232547419 +0000 UTC m=+1117.316147846" watchObservedRunningTime="2025-09-30 14:38:15.24488774 +0000 UTC m=+1117.328488187" Sep 30 14:38:16 crc kubenswrapper[4799]: I0930 14:38:16.191352 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:16 crc kubenswrapper[4799]: E0930 14:38:16.192020 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:38:16 crc kubenswrapper[4799]: E0930 14:38:16.192061 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:38:16 crc kubenswrapper[4799]: E0930 14:38:16.192143 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift podName:1500572d-64d4-483a-a525-bc0530fe4cbb nodeName:}" failed. No retries permitted until 2025-09-30 14:38:20.192116651 +0000 UTC m=+1122.275717088 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift") pod "swift-storage-0" (UID: "1500572d-64d4-483a-a525-bc0530fe4cbb") : configmap "swift-ring-files" not found Sep 30 14:38:16 crc kubenswrapper[4799]: I0930 14:38:16.231308 4799 generic.go:334] "Generic (PLEG): container finished" podID="4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b" containerID="34e9f5d1a8a89605dc771a739f141bfb3191b1f19adba7d51b71aa26eae4c0ab" exitCode=0 Sep 30 14:38:16 crc kubenswrapper[4799]: I0930 14:38:16.231413 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kv6wf" event={"ID":"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b","Type":"ContainerDied","Data":"34e9f5d1a8a89605dc771a739f141bfb3191b1f19adba7d51b71aa26eae4c0ab"} Sep 30 14:38:17 crc kubenswrapper[4799]: I0930 14:38:17.988781 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kv6wf" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.125033 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48plr\" (UniqueName: \"kubernetes.io/projected/4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b-kube-api-access-48plr\") pod \"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b\" (UID: \"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b\") " Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.133934 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b-kube-api-access-48plr" (OuterVolumeSpecName: "kube-api-access-48plr") pod "4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b" (UID: "4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b"). InnerVolumeSpecName "kube-api-access-48plr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.223993 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b430-account-create-bh8t8"] Sep 30 14:38:18 crc kubenswrapper[4799]: E0930 14:38:18.224418 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b" containerName="mariadb-database-create" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.224436 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b" containerName="mariadb-database-create" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.224599 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b" containerName="mariadb-database-create" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.225219 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b430-account-create-bh8t8" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.226663 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fktbx\" (UniqueName: \"kubernetes.io/projected/d63ec7b1-a806-40c4-9f60-a17b892e27cf-kube-api-access-fktbx\") pod \"keystone-b430-account-create-bh8t8\" (UID: \"d63ec7b1-a806-40c4-9f60-a17b892e27cf\") " pod="openstack/keystone-b430-account-create-bh8t8" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.226931 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48plr\" (UniqueName: \"kubernetes.io/projected/4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b-kube-api-access-48plr\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.228923 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.248135 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b430-account-create-bh8t8"] Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.255391 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kv6wf" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.256781 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kv6wf" event={"ID":"4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b","Type":"ContainerDied","Data":"cf9464b3dde9304cbec22f746b905fb61962b3ad1936005da38b4d5c0089b5d9"} Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.256847 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf9464b3dde9304cbec22f746b905fb61962b3ad1936005da38b4d5c0089b5d9" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.263880 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mbljj" event={"ID":"ad219d29-0576-4b77-b5b9-fc2c980b7093","Type":"ContainerStarted","Data":"6db754fbea38bc697cbd3d89d4a1f6dd5041b1b72a7cd401ae5ae357dacf44c2"} Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.295857 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-mbljj" podStartSLOduration=2.15158203 podStartE2EDuration="6.295832402s" podCreationTimestamp="2025-09-30 14:38:12 +0000 UTC" firstStartedPulling="2025-09-30 14:38:13.830166753 +0000 UTC m=+1115.913767180" lastFinishedPulling="2025-09-30 14:38:17.974417125 +0000 UTC m=+1120.058017552" observedRunningTime="2025-09-30 14:38:18.293931386 +0000 UTC m=+1120.377531813" watchObservedRunningTime="2025-09-30 14:38:18.295832402 +0000 UTC m=+1120.379432839" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.328736 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fktbx\" (UniqueName: \"kubernetes.io/projected/d63ec7b1-a806-40c4-9f60-a17b892e27cf-kube-api-access-fktbx\") pod \"keystone-b430-account-create-bh8t8\" (UID: \"d63ec7b1-a806-40c4-9f60-a17b892e27cf\") " pod="openstack/keystone-b430-account-create-bh8t8" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.348522 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fktbx\" (UniqueName: \"kubernetes.io/projected/d63ec7b1-a806-40c4-9f60-a17b892e27cf-kube-api-access-fktbx\") pod \"keystone-b430-account-create-bh8t8\" (UID: \"d63ec7b1-a806-40c4-9f60-a17b892e27cf\") " pod="openstack/keystone-b430-account-create-bh8t8" Sep 30 14:38:18 crc kubenswrapper[4799]: I0930 14:38:18.553938 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b430-account-create-bh8t8" Sep 30 14:38:19 crc kubenswrapper[4799]: I0930 14:38:19.008099 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b430-account-create-bh8t8"] Sep 30 14:38:19 crc kubenswrapper[4799]: W0930 14:38:19.013110 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd63ec7b1_a806_40c4_9f60_a17b892e27cf.slice/crio-792b9cea0133a159097b2aae19404d1ac138ddda29e3102ba39112c39332874b WatchSource:0}: Error finding container 792b9cea0133a159097b2aae19404d1ac138ddda29e3102ba39112c39332874b: Status 404 returned error can't find the container with id 792b9cea0133a159097b2aae19404d1ac138ddda29e3102ba39112c39332874b Sep 30 14:38:19 crc kubenswrapper[4799]: I0930 14:38:19.278972 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b430-account-create-bh8t8" event={"ID":"d63ec7b1-a806-40c4-9f60-a17b892e27cf","Type":"ContainerStarted","Data":"9180d755716119ac42d286c0e8752d52e15d1321c059368d6fd7c955d3700d5c"} Sep 30 14:38:19 crc kubenswrapper[4799]: I0930 14:38:19.279354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b430-account-create-bh8t8" event={"ID":"d63ec7b1-a806-40c4-9f60-a17b892e27cf","Type":"ContainerStarted","Data":"792b9cea0133a159097b2aae19404d1ac138ddda29e3102ba39112c39332874b"} Sep 30 14:38:19 crc kubenswrapper[4799]: I0930 14:38:19.300239 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-b430-account-create-bh8t8" podStartSLOduration=1.300211203 podStartE2EDuration="1.300211203s" podCreationTimestamp="2025-09-30 14:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:38:19.299545403 +0000 UTC m=+1121.383145840" watchObservedRunningTime="2025-09-30 14:38:19.300211203 +0000 UTC m=+1121.383811640" Sep 30 14:38:20 crc kubenswrapper[4799]: I0930 14:38:20.273372 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:20 crc kubenswrapper[4799]: E0930 14:38:20.273696 4799 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Sep 30 14:38:20 crc kubenswrapper[4799]: E0930 14:38:20.273735 4799 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Sep 30 14:38:20 crc kubenswrapper[4799]: E0930 14:38:20.273822 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift podName:1500572d-64d4-483a-a525-bc0530fe4cbb nodeName:}" failed. No retries permitted until 2025-09-30 14:38:28.273792884 +0000 UTC m=+1130.357393311 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift") pod "swift-storage-0" (UID: "1500572d-64d4-483a-a525-bc0530fe4cbb") : configmap "swift-ring-files" not found Sep 30 14:38:20 crc kubenswrapper[4799]: I0930 14:38:20.286291 4799 generic.go:334] "Generic (PLEG): container finished" podID="d63ec7b1-a806-40c4-9f60-a17b892e27cf" containerID="9180d755716119ac42d286c0e8752d52e15d1321c059368d6fd7c955d3700d5c" exitCode=0 Sep 30 14:38:20 crc kubenswrapper[4799]: I0930 14:38:20.286358 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b430-account-create-bh8t8" event={"ID":"d63ec7b1-a806-40c4-9f60-a17b892e27cf","Type":"ContainerDied","Data":"9180d755716119ac42d286c0e8752d52e15d1321c059368d6fd7c955d3700d5c"} Sep 30 14:38:21 crc kubenswrapper[4799]: I0930 14:38:21.506937 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:38:21 crc kubenswrapper[4799]: I0930 14:38:21.598265 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9jrdq"] Sep 30 14:38:21 crc kubenswrapper[4799]: I0930 14:38:21.598625 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" podUID="b653ad5b-ace8-4659-8f9c-af1bb901776f" containerName="dnsmasq-dns" containerID="cri-o://ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b" gracePeriod=10 Sep 30 14:38:21 crc kubenswrapper[4799]: I0930 14:38:21.821147 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b430-account-create-bh8t8" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.003176 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fktbx\" (UniqueName: \"kubernetes.io/projected/d63ec7b1-a806-40c4-9f60-a17b892e27cf-kube-api-access-fktbx\") pod \"d63ec7b1-a806-40c4-9f60-a17b892e27cf\" (UID: \"d63ec7b1-a806-40c4-9f60-a17b892e27cf\") " Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.010194 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d63ec7b1-a806-40c4-9f60-a17b892e27cf-kube-api-access-fktbx" (OuterVolumeSpecName: "kube-api-access-fktbx") pod "d63ec7b1-a806-40c4-9f60-a17b892e27cf" (UID: "d63ec7b1-a806-40c4-9f60-a17b892e27cf"). InnerVolumeSpecName "kube-api-access-fktbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.105995 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fktbx\" (UniqueName: \"kubernetes.io/projected/d63ec7b1-a806-40c4-9f60-a17b892e27cf-kube-api-access-fktbx\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.217578 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.305207 4799 generic.go:334] "Generic (PLEG): container finished" podID="b653ad5b-ace8-4659-8f9c-af1bb901776f" containerID="ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b" exitCode=0 Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.305285 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.305314 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" event={"ID":"b653ad5b-ace8-4659-8f9c-af1bb901776f","Type":"ContainerDied","Data":"ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b"} Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.305384 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9jrdq" event={"ID":"b653ad5b-ace8-4659-8f9c-af1bb901776f","Type":"ContainerDied","Data":"2db3943b54d96110da96e33ff95aa78e8fe0ee950c81a42257e92745f7cded6f"} Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.305414 4799 scope.go:117] "RemoveContainer" containerID="ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.307980 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-dns-svc\") pod \"b653ad5b-ace8-4659-8f9c-af1bb901776f\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.308057 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ks2h\" (UniqueName: \"kubernetes.io/projected/b653ad5b-ace8-4659-8f9c-af1bb901776f-kube-api-access-2ks2h\") pod \"b653ad5b-ace8-4659-8f9c-af1bb901776f\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.308927 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-config\") pod \"b653ad5b-ace8-4659-8f9c-af1bb901776f\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.308968 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-nb\") pod \"b653ad5b-ace8-4659-8f9c-af1bb901776f\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.309001 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-sb\") pod \"b653ad5b-ace8-4659-8f9c-af1bb901776f\" (UID: \"b653ad5b-ace8-4659-8f9c-af1bb901776f\") " Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.312153 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b430-account-create-bh8t8" event={"ID":"d63ec7b1-a806-40c4-9f60-a17b892e27cf","Type":"ContainerDied","Data":"792b9cea0133a159097b2aae19404d1ac138ddda29e3102ba39112c39332874b"} Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.312210 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="792b9cea0133a159097b2aae19404d1ac138ddda29e3102ba39112c39332874b" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.312322 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b430-account-create-bh8t8" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.325220 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b653ad5b-ace8-4659-8f9c-af1bb901776f-kube-api-access-2ks2h" (OuterVolumeSpecName: "kube-api-access-2ks2h") pod "b653ad5b-ace8-4659-8f9c-af1bb901776f" (UID: "b653ad5b-ace8-4659-8f9c-af1bb901776f"). InnerVolumeSpecName "kube-api-access-2ks2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.344131 4799 scope.go:117] "RemoveContainer" containerID="3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.392962 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b653ad5b-ace8-4659-8f9c-af1bb901776f" (UID: "b653ad5b-ace8-4659-8f9c-af1bb901776f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.394448 4799 scope.go:117] "RemoveContainer" containerID="ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b" Sep 30 14:38:22 crc kubenswrapper[4799]: E0930 14:38:22.395344 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b\": container with ID starting with ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b not found: ID does not exist" containerID="ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.395397 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b"} err="failed to get container status \"ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b\": rpc error: code = NotFound desc = could not find container \"ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b\": container with ID starting with ae4f7dc7ec3d1fe714d0510c725ab59b23628da2de9eb4b6edc9d0350824b30b not found: ID does not exist" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.395423 4799 scope.go:117] "RemoveContainer" containerID="3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.396872 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b653ad5b-ace8-4659-8f9c-af1bb901776f" (UID: "b653ad5b-ace8-4659-8f9c-af1bb901776f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:22 crc kubenswrapper[4799]: E0930 14:38:22.399966 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09\": container with ID starting with 3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09 not found: ID does not exist" containerID="3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.400017 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09"} err="failed to get container status \"3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09\": rpc error: code = NotFound desc = could not find container \"3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09\": container with ID starting with 3bbab70bda7f177e6a883ec7b6aa88e47bd25802e7aca84675b9eee6900bfa09 not found: ID does not exist" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.400384 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-config" (OuterVolumeSpecName: "config") pod "b653ad5b-ace8-4659-8f9c-af1bb901776f" (UID: "b653ad5b-ace8-4659-8f9c-af1bb901776f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.411191 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.411230 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.411246 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.411259 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ks2h\" (UniqueName: \"kubernetes.io/projected/b653ad5b-ace8-4659-8f9c-af1bb901776f-kube-api-access-2ks2h\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.413069 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b653ad5b-ace8-4659-8f9c-af1bb901776f" (UID: "b653ad5b-ace8-4659-8f9c-af1bb901776f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.520808 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b653ad5b-ace8-4659-8f9c-af1bb901776f-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.627129 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9jrdq"] Sep 30 14:38:22 crc kubenswrapper[4799]: I0930 14:38:22.638275 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9jrdq"] Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.102676 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-9117-account-create-qfjcx"] Sep 30 14:38:24 crc kubenswrapper[4799]: E0930 14:38:24.103330 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63ec7b1-a806-40c4-9f60-a17b892e27cf" containerName="mariadb-account-create" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.103345 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63ec7b1-a806-40c4-9f60-a17b892e27cf" containerName="mariadb-account-create" Sep 30 14:38:24 crc kubenswrapper[4799]: E0930 14:38:24.103381 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b653ad5b-ace8-4659-8f9c-af1bb901776f" containerName="init" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.103387 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b653ad5b-ace8-4659-8f9c-af1bb901776f" containerName="init" Sep 30 14:38:24 crc kubenswrapper[4799]: E0930 14:38:24.103395 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b653ad5b-ace8-4659-8f9c-af1bb901776f" containerName="dnsmasq-dns" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.103402 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b653ad5b-ace8-4659-8f9c-af1bb901776f" containerName="dnsmasq-dns" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.103611 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63ec7b1-a806-40c4-9f60-a17b892e27cf" containerName="mariadb-account-create" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.103628 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b653ad5b-ace8-4659-8f9c-af1bb901776f" containerName="dnsmasq-dns" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.104374 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9117-account-create-qfjcx" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.106891 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.121501 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9117-account-create-qfjcx"] Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.248786 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw927\" (UniqueName: \"kubernetes.io/projected/a0d8d726-9fbf-4a54-bcd1-835dc062709f-kube-api-access-hw927\") pod \"glance-9117-account-create-qfjcx\" (UID: \"a0d8d726-9fbf-4a54-bcd1-835dc062709f\") " pod="openstack/glance-9117-account-create-qfjcx" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.351285 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw927\" (UniqueName: \"kubernetes.io/projected/a0d8d726-9fbf-4a54-bcd1-835dc062709f-kube-api-access-hw927\") pod \"glance-9117-account-create-qfjcx\" (UID: \"a0d8d726-9fbf-4a54-bcd1-835dc062709f\") " pod="openstack/glance-9117-account-create-qfjcx" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.378121 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw927\" (UniqueName: \"kubernetes.io/projected/a0d8d726-9fbf-4a54-bcd1-835dc062709f-kube-api-access-hw927\") pod \"glance-9117-account-create-qfjcx\" (UID: \"a0d8d726-9fbf-4a54-bcd1-835dc062709f\") " pod="openstack/glance-9117-account-create-qfjcx" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.433961 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9117-account-create-qfjcx" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.517733 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b653ad5b-ace8-4659-8f9c-af1bb901776f" path="/var/lib/kubelet/pods/b653ad5b-ace8-4659-8f9c-af1bb901776f/volumes" Sep 30 14:38:24 crc kubenswrapper[4799]: I0930 14:38:24.752704 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9117-account-create-qfjcx"] Sep 30 14:38:25 crc kubenswrapper[4799]: I0930 14:38:25.283568 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Sep 30 14:38:25 crc kubenswrapper[4799]: I0930 14:38:25.343745 4799 generic.go:334] "Generic (PLEG): container finished" podID="a0d8d726-9fbf-4a54-bcd1-835dc062709f" containerID="a70bfcac89f1ba8547ee90d38bc9a91f00495ceea0a895028b8bec5c8e914cfe" exitCode=0 Sep 30 14:38:25 crc kubenswrapper[4799]: I0930 14:38:25.344321 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9117-account-create-qfjcx" event={"ID":"a0d8d726-9fbf-4a54-bcd1-835dc062709f","Type":"ContainerDied","Data":"a70bfcac89f1ba8547ee90d38bc9a91f00495ceea0a895028b8bec5c8e914cfe"} Sep 30 14:38:25 crc kubenswrapper[4799]: I0930 14:38:25.344610 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9117-account-create-qfjcx" event={"ID":"a0d8d726-9fbf-4a54-bcd1-835dc062709f","Type":"ContainerStarted","Data":"863fca843c2b7676f29d3807b0e163d661a558d6223fcfb0d9b2f76c95a8be86"} Sep 30 14:38:26 crc kubenswrapper[4799]: I0930 14:38:26.682646 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9117-account-create-qfjcx" Sep 30 14:38:26 crc kubenswrapper[4799]: I0930 14:38:26.703099 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw927\" (UniqueName: \"kubernetes.io/projected/a0d8d726-9fbf-4a54-bcd1-835dc062709f-kube-api-access-hw927\") pod \"a0d8d726-9fbf-4a54-bcd1-835dc062709f\" (UID: \"a0d8d726-9fbf-4a54-bcd1-835dc062709f\") " Sep 30 14:38:26 crc kubenswrapper[4799]: I0930 14:38:26.720437 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0d8d726-9fbf-4a54-bcd1-835dc062709f-kube-api-access-hw927" (OuterVolumeSpecName: "kube-api-access-hw927") pod "a0d8d726-9fbf-4a54-bcd1-835dc062709f" (UID: "a0d8d726-9fbf-4a54-bcd1-835dc062709f"). InnerVolumeSpecName "kube-api-access-hw927". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:26 crc kubenswrapper[4799]: I0930 14:38:26.805621 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw927\" (UniqueName: \"kubernetes.io/projected/a0d8d726-9fbf-4a54-bcd1-835dc062709f-kube-api-access-hw927\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:27 crc kubenswrapper[4799]: I0930 14:38:27.362192 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9117-account-create-qfjcx" Sep 30 14:38:27 crc kubenswrapper[4799]: I0930 14:38:27.362295 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9117-account-create-qfjcx" event={"ID":"a0d8d726-9fbf-4a54-bcd1-835dc062709f","Type":"ContainerDied","Data":"863fca843c2b7676f29d3807b0e163d661a558d6223fcfb0d9b2f76c95a8be86"} Sep 30 14:38:27 crc kubenswrapper[4799]: I0930 14:38:27.362377 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="863fca843c2b7676f29d3807b0e163d661a558d6223fcfb0d9b2f76c95a8be86" Sep 30 14:38:27 crc kubenswrapper[4799]: I0930 14:38:27.365131 4799 generic.go:334] "Generic (PLEG): container finished" podID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerID="0817b5a4428f559089509977d4a81d1d3084c69e64b252f2427d59eae955efa7" exitCode=0 Sep 30 14:38:27 crc kubenswrapper[4799]: I0930 14:38:27.365283 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb82f13-0450-42a8-9c8a-0c174c534382","Type":"ContainerDied","Data":"0817b5a4428f559089509977d4a81d1d3084c69e64b252f2427d59eae955efa7"} Sep 30 14:38:27 crc kubenswrapper[4799]: I0930 14:38:27.366998 4799 generic.go:334] "Generic (PLEG): container finished" podID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerID="4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3" exitCode=0 Sep 30 14:38:27 crc kubenswrapper[4799]: I0930 14:38:27.367046 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef6dcf7e-70a0-4759-afef-53e70b800ba0","Type":"ContainerDied","Data":"4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3"} Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.335060 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.352622 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1500572d-64d4-483a-a525-bc0530fe4cbb-etc-swift\") pod \"swift-storage-0\" (UID: \"1500572d-64d4-483a-a525-bc0530fe4cbb\") " pod="openstack/swift-storage-0" Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.378432 4799 generic.go:334] "Generic (PLEG): container finished" podID="ad219d29-0576-4b77-b5b9-fc2c980b7093" containerID="6db754fbea38bc697cbd3d89d4a1f6dd5041b1b72a7cd401ae5ae357dacf44c2" exitCode=0 Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.378530 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mbljj" event={"ID":"ad219d29-0576-4b77-b5b9-fc2c980b7093","Type":"ContainerDied","Data":"6db754fbea38bc697cbd3d89d4a1f6dd5041b1b72a7cd401ae5ae357dacf44c2"} Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.383020 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb82f13-0450-42a8-9c8a-0c174c534382","Type":"ContainerStarted","Data":"d0a099467cee71f3c37c55698919e290d84a71960fad48a65565f735a6ca095f"} Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.383415 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.386235 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef6dcf7e-70a0-4759-afef-53e70b800ba0","Type":"ContainerStarted","Data":"d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7"} Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.386538 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.401302 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Sep 30 14:38:28 crc kubenswrapper[4799]: I0930 14:38:28.450750 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371951.404058 podStartE2EDuration="1m25.450717348s" podCreationTimestamp="2025-09-30 14:37:03 +0000 UTC" firstStartedPulling="2025-09-30 14:37:05.825746863 +0000 UTC m=+1047.909347290" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:38:28.448683658 +0000 UTC m=+1130.532284105" watchObservedRunningTime="2025-09-30 14:38:28.450717348 +0000 UTC m=+1130.534317775" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.041038 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.474247538 podStartE2EDuration="1m26.041003084s" podCreationTimestamp="2025-09-30 14:37:03 +0000 UTC" firstStartedPulling="2025-09-30 14:37:05.450167653 +0000 UTC m=+1047.533768080" lastFinishedPulling="2025-09-30 14:37:56.016923189 +0000 UTC m=+1098.100523626" observedRunningTime="2025-09-30 14:38:28.507989242 +0000 UTC m=+1130.591589679" watchObservedRunningTime="2025-09-30 14:38:29.041003084 +0000 UTC m=+1131.124603511" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.044686 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8a5a-account-create-kmzh9"] Sep 30 14:38:29 crc kubenswrapper[4799]: E0930 14:38:29.045201 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d8d726-9fbf-4a54-bcd1-835dc062709f" containerName="mariadb-account-create" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.045221 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d8d726-9fbf-4a54-bcd1-835dc062709f" containerName="mariadb-account-create" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.045426 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d8d726-9fbf-4a54-bcd1-835dc062709f" containerName="mariadb-account-create" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.046228 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8a5a-account-create-kmzh9" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.049877 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.079959 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8a5a-account-create-kmzh9"] Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.093165 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Sep 30 14:38:29 crc kubenswrapper[4799]: W0930 14:38:29.112985 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1500572d_64d4_483a_a525_bc0530fe4cbb.slice/crio-67f22836c514c2da0de423be0ea3cbb7c315218f1a31d80bc88d0e53a7bfcee1 WatchSource:0}: Error finding container 67f22836c514c2da0de423be0ea3cbb7c315218f1a31d80bc88d0e53a7bfcee1: Status 404 returned error can't find the container with id 67f22836c514c2da0de423be0ea3cbb7c315218f1a31d80bc88d0e53a7bfcee1 Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.149185 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmbv6\" (UniqueName: \"kubernetes.io/projected/de063760-02b1-4dae-b475-8e33b7f5bf66-kube-api-access-cmbv6\") pod \"placement-8a5a-account-create-kmzh9\" (UID: \"de063760-02b1-4dae-b475-8e33b7f5bf66\") " pod="openstack/placement-8a5a-account-create-kmzh9" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.250359 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmbv6\" (UniqueName: \"kubernetes.io/projected/de063760-02b1-4dae-b475-8e33b7f5bf66-kube-api-access-cmbv6\") pod \"placement-8a5a-account-create-kmzh9\" (UID: \"de063760-02b1-4dae-b475-8e33b7f5bf66\") " pod="openstack/placement-8a5a-account-create-kmzh9" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.281419 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmbv6\" (UniqueName: \"kubernetes.io/projected/de063760-02b1-4dae-b475-8e33b7f5bf66-kube-api-access-cmbv6\") pod \"placement-8a5a-account-create-kmzh9\" (UID: \"de063760-02b1-4dae-b475-8e33b7f5bf66\") " pod="openstack/placement-8a5a-account-create-kmzh9" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.347434 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4d9n4"] Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.349411 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.361328 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.361802 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pjthr" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.381067 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8a5a-account-create-kmzh9" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.384296 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4d9n4"] Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.413985 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"67f22836c514c2da0de423be0ea3cbb7c315218f1a31d80bc88d0e53a7bfcee1"} Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.454781 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4kwk\" (UniqueName: \"kubernetes.io/projected/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-kube-api-access-n4kwk\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.455001 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-combined-ca-bundle\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.455086 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-db-sync-config-data\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.455216 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-config-data\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.560641 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-combined-ca-bundle\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.561512 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-db-sync-config-data\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.561709 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-config-data\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.561910 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4kwk\" (UniqueName: \"kubernetes.io/projected/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-kube-api-access-n4kwk\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.569319 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-config-data\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.572147 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-db-sync-config-data\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.578358 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-combined-ca-bundle\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.617383 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4kwk\" (UniqueName: \"kubernetes.io/projected/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-kube-api-access-n4kwk\") pod \"glance-db-sync-4d9n4\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.649606 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.649695 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.649761 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.650561 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c08cc75dc1edb21f93545a035f6f554f287534552d52a3400113832177d88167"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.650615 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://c08cc75dc1edb21f93545a035f6f554f287534552d52a3400113832177d88167" gracePeriod=600 Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.678425 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4d9n4" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.862297 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.981012 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-dispersionconf\") pod \"ad219d29-0576-4b77-b5b9-fc2c980b7093\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.981800 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-combined-ca-bundle\") pod \"ad219d29-0576-4b77-b5b9-fc2c980b7093\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.982115 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmtvm\" (UniqueName: \"kubernetes.io/projected/ad219d29-0576-4b77-b5b9-fc2c980b7093-kube-api-access-qmtvm\") pod \"ad219d29-0576-4b77-b5b9-fc2c980b7093\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.982166 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-ring-data-devices\") pod \"ad219d29-0576-4b77-b5b9-fc2c980b7093\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.982213 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-scripts\") pod \"ad219d29-0576-4b77-b5b9-fc2c980b7093\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.982369 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-swiftconf\") pod \"ad219d29-0576-4b77-b5b9-fc2c980b7093\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.982438 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad219d29-0576-4b77-b5b9-fc2c980b7093-etc-swift\") pod \"ad219d29-0576-4b77-b5b9-fc2c980b7093\" (UID: \"ad219d29-0576-4b77-b5b9-fc2c980b7093\") " Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.985456 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ad219d29-0576-4b77-b5b9-fc2c980b7093" (UID: "ad219d29-0576-4b77-b5b9-fc2c980b7093"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.986944 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad219d29-0576-4b77-b5b9-fc2c980b7093-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ad219d29-0576-4b77-b5b9-fc2c980b7093" (UID: "ad219d29-0576-4b77-b5b9-fc2c980b7093"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.988408 4799 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad219d29-0576-4b77-b5b9-fc2c980b7093-etc-swift\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.988446 4799 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-ring-data-devices\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:29 crc kubenswrapper[4799]: I0930 14:38:29.992111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad219d29-0576-4b77-b5b9-fc2c980b7093-kube-api-access-qmtvm" (OuterVolumeSpecName: "kube-api-access-qmtvm") pod "ad219d29-0576-4b77-b5b9-fc2c980b7093" (UID: "ad219d29-0576-4b77-b5b9-fc2c980b7093"). InnerVolumeSpecName "kube-api-access-qmtvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.010385 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ad219d29-0576-4b77-b5b9-fc2c980b7093" (UID: "ad219d29-0576-4b77-b5b9-fc2c980b7093"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.050211 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad219d29-0576-4b77-b5b9-fc2c980b7093" (UID: "ad219d29-0576-4b77-b5b9-fc2c980b7093"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.085364 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ad219d29-0576-4b77-b5b9-fc2c980b7093" (UID: "ad219d29-0576-4b77-b5b9-fc2c980b7093"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.092822 4799 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-swiftconf\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.092899 4799 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-dispersionconf\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.092915 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad219d29-0576-4b77-b5b9-fc2c980b7093-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.092933 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmtvm\" (UniqueName: \"kubernetes.io/projected/ad219d29-0576-4b77-b5b9-fc2c980b7093-kube-api-access-qmtvm\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.098152 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-scripts" (OuterVolumeSpecName: "scripts") pod "ad219d29-0576-4b77-b5b9-fc2c980b7093" (UID: "ad219d29-0576-4b77-b5b9-fc2c980b7093"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.194132 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad219d29-0576-4b77-b5b9-fc2c980b7093-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.211768 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8a5a-account-create-kmzh9"] Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.427704 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="c08cc75dc1edb21f93545a035f6f554f287534552d52a3400113832177d88167" exitCode=0 Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.428086 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"c08cc75dc1edb21f93545a035f6f554f287534552d52a3400113832177d88167"} Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.428925 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"bb9ad6936dc7236258e4efd5f74222bbe45a21ab72fed8d7b84c8d34eafe641b"} Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.428948 4799 scope.go:117] "RemoveContainer" containerID="7be6e12e797fed54fc7aab60eaeaf0c5516974a723cfc906b94db6e0a39463b8" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.444583 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8a5a-account-create-kmzh9" event={"ID":"de063760-02b1-4dae-b475-8e33b7f5bf66","Type":"ContainerStarted","Data":"8908da74cef5d64ed919f1a1ab55be1b841a3a18b230221ad61eb952cfc23d67"} Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.452609 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mbljj" event={"ID":"ad219d29-0576-4b77-b5b9-fc2c980b7093","Type":"ContainerDied","Data":"8c1f7b80c591f62727a45edeb472f2b4f4b10a6fd2e633bd8cd0c34d034e480d"} Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.452694 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c1f7b80c591f62727a45edeb472f2b4f4b10a6fd2e633bd8cd0c34d034e480d" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.452789 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mbljj" Sep 30 14:38:30 crc kubenswrapper[4799]: I0930 14:38:30.579568 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4d9n4"] Sep 30 14:38:31 crc kubenswrapper[4799]: I0930 14:38:31.470261 4799 generic.go:334] "Generic (PLEG): container finished" podID="de063760-02b1-4dae-b475-8e33b7f5bf66" containerID="987166546ed56ca282df48dcdde95734d12003ca61d52d68a574dd3ee10b92fb" exitCode=0 Sep 30 14:38:31 crc kubenswrapper[4799]: I0930 14:38:31.470356 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8a5a-account-create-kmzh9" event={"ID":"de063760-02b1-4dae-b475-8e33b7f5bf66","Type":"ContainerDied","Data":"987166546ed56ca282df48dcdde95734d12003ca61d52d68a574dd3ee10b92fb"} Sep 30 14:38:31 crc kubenswrapper[4799]: I0930 14:38:31.472290 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4d9n4" event={"ID":"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b","Type":"ContainerStarted","Data":"3313cf6d4f81f4ac8c2ad5faddff4fb04ba86817410812a0c510e52b218de228"} Sep 30 14:38:32 crc kubenswrapper[4799]: I0930 14:38:32.484576 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"3502d4b275749b55aaa28a5e9372e9695d01a1f9d3384075f84fcdf100581072"} Sep 30 14:38:32 crc kubenswrapper[4799]: I0930 14:38:32.484942 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"c0abd0d412883c93b2d881f8b0b28415ab07ac4f26db8e390457941775b9055b"} Sep 30 14:38:32 crc kubenswrapper[4799]: I0930 14:38:32.484953 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"60b6551d484c3945549a45baad1a252e2660997017c144c1f521b94af78f3dd6"} Sep 30 14:38:32 crc kubenswrapper[4799]: I0930 14:38:32.484962 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"a188b9ce186caebe0eb5a5af401fa841b22504dc0d1deb835c56074233b42ec8"} Sep 30 14:38:32 crc kubenswrapper[4799]: I0930 14:38:32.562951 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2kvml" podUID="cbefa245-5ad5-4303-b166-59574e8a0c65" containerName="ovn-controller" probeResult="failure" output=< Sep 30 14:38:32 crc kubenswrapper[4799]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 14:38:32 crc kubenswrapper[4799]: > Sep 30 14:38:32 crc kubenswrapper[4799]: I0930 14:38:32.991313 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8a5a-account-create-kmzh9" Sep 30 14:38:33 crc kubenswrapper[4799]: I0930 14:38:33.051812 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmbv6\" (UniqueName: \"kubernetes.io/projected/de063760-02b1-4dae-b475-8e33b7f5bf66-kube-api-access-cmbv6\") pod \"de063760-02b1-4dae-b475-8e33b7f5bf66\" (UID: \"de063760-02b1-4dae-b475-8e33b7f5bf66\") " Sep 30 14:38:33 crc kubenswrapper[4799]: I0930 14:38:33.075990 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de063760-02b1-4dae-b475-8e33b7f5bf66-kube-api-access-cmbv6" (OuterVolumeSpecName: "kube-api-access-cmbv6") pod "de063760-02b1-4dae-b475-8e33b7f5bf66" (UID: "de063760-02b1-4dae-b475-8e33b7f5bf66"). InnerVolumeSpecName "kube-api-access-cmbv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:33 crc kubenswrapper[4799]: I0930 14:38:33.154517 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmbv6\" (UniqueName: \"kubernetes.io/projected/de063760-02b1-4dae-b475-8e33b7f5bf66-kube-api-access-cmbv6\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:33 crc kubenswrapper[4799]: I0930 14:38:33.503310 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8a5a-account-create-kmzh9" event={"ID":"de063760-02b1-4dae-b475-8e33b7f5bf66","Type":"ContainerDied","Data":"8908da74cef5d64ed919f1a1ab55be1b841a3a18b230221ad61eb952cfc23d67"} Sep 30 14:38:33 crc kubenswrapper[4799]: I0930 14:38:33.503358 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8908da74cef5d64ed919f1a1ab55be1b841a3a18b230221ad61eb952cfc23d67" Sep 30 14:38:33 crc kubenswrapper[4799]: I0930 14:38:33.503404 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8a5a-account-create-kmzh9" Sep 30 14:38:36 crc kubenswrapper[4799]: I0930 14:38:36.538178 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"a71bccce0febe61214ca07f8373392c49fff628f780d90b2fc69629b270a3f08"} Sep 30 14:38:36 crc kubenswrapper[4799]: I0930 14:38:36.539751 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"2ded521bae3dc7f1220fbf72ec9e6da8069c9c7758c0d66a24abc9633de3368f"} Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.561567 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"9ed7c61814f3c1f85426c261c32537e17bf212c9fa9d9d2df51c0906aa995d1b"} Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.561624 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"af0b942928bbdf7396e4a79e3025a00d81ebb8895b3e91b056d660d23a3751ab"} Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.567472 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2kvml" podUID="cbefa245-5ad5-4303-b166-59574e8a0c65" containerName="ovn-controller" probeResult="failure" output=< Sep 30 14:38:37 crc kubenswrapper[4799]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 14:38:37 crc kubenswrapper[4799]: > Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.622145 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.625041 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v2rs4" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.883004 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2kvml-config-4h5lr"] Sep 30 14:38:37 crc kubenswrapper[4799]: E0930 14:38:37.885472 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de063760-02b1-4dae-b475-8e33b7f5bf66" containerName="mariadb-account-create" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.885549 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="de063760-02b1-4dae-b475-8e33b7f5bf66" containerName="mariadb-account-create" Sep 30 14:38:37 crc kubenswrapper[4799]: E0930 14:38:37.885622 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad219d29-0576-4b77-b5b9-fc2c980b7093" containerName="swift-ring-rebalance" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.885710 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad219d29-0576-4b77-b5b9-fc2c980b7093" containerName="swift-ring-rebalance" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.886060 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="de063760-02b1-4dae-b475-8e33b7f5bf66" containerName="mariadb-account-create" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.886326 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad219d29-0576-4b77-b5b9-fc2c980b7093" containerName="swift-ring-rebalance" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.887037 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.893846 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.901926 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2kvml-config-4h5lr"] Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.946049 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run-ovn\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.946418 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-log-ovn\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.946527 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-additional-scripts\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.946622 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-scripts\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.946755 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stb26\" (UniqueName: \"kubernetes.io/projected/09c40c7c-8032-4c68-b038-b427b089a157-kube-api-access-stb26\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:37 crc kubenswrapper[4799]: I0930 14:38:37.946844 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.048183 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-additional-scripts\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.048250 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-scripts\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.048309 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stb26\" (UniqueName: \"kubernetes.io/projected/09c40c7c-8032-4c68-b038-b427b089a157-kube-api-access-stb26\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.048336 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.048399 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run-ovn\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.048458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-log-ovn\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.048852 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-log-ovn\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.049677 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-additional-scripts\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.051479 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run-ovn\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.051479 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.053480 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-scripts\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.073783 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stb26\" (UniqueName: \"kubernetes.io/projected/09c40c7c-8032-4c68-b038-b427b089a157-kube-api-access-stb26\") pod \"ovn-controller-2kvml-config-4h5lr\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:38 crc kubenswrapper[4799]: I0930 14:38:38.217307 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:42 crc kubenswrapper[4799]: I0930 14:38:42.566131 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2kvml" podUID="cbefa245-5ad5-4303-b166-59574e8a0c65" containerName="ovn-controller" probeResult="failure" output=< Sep 30 14:38:42 crc kubenswrapper[4799]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 14:38:42 crc kubenswrapper[4799]: > Sep 30 14:38:44 crc kubenswrapper[4799]: I0930 14:38:44.725986 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:38:45 crc kubenswrapper[4799]: I0930 14:38:45.135437 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 30 14:38:46 crc kubenswrapper[4799]: I0930 14:38:46.962191 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2kvml-config-4h5lr"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.056256 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-7gkqb"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.057667 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7gkqb" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.081789 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-7gkqb"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.240994 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-452wh\" (UniqueName: \"kubernetes.io/projected/8bb77af6-6c03-44f8-bec8-d168840a0c1e-kube-api-access-452wh\") pod \"cinder-db-create-7gkqb\" (UID: \"8bb77af6-6c03-44f8-bec8-d168840a0c1e\") " pod="openstack/cinder-db-create-7gkqb" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.272729 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-m88q9"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.293946 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m88q9" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.308203 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m88q9"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.344469 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-452wh\" (UniqueName: \"kubernetes.io/projected/8bb77af6-6c03-44f8-bec8-d168840a0c1e-kube-api-access-452wh\") pod \"cinder-db-create-7gkqb\" (UID: \"8bb77af6-6c03-44f8-bec8-d168840a0c1e\") " pod="openstack/cinder-db-create-7gkqb" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.375683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-452wh\" (UniqueName: \"kubernetes.io/projected/8bb77af6-6c03-44f8-bec8-d168840a0c1e-kube-api-access-452wh\") pod \"cinder-db-create-7gkqb\" (UID: \"8bb77af6-6c03-44f8-bec8-d168840a0c1e\") " pod="openstack/cinder-db-create-7gkqb" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.394987 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7gkqb" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.449070 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5dng\" (UniqueName: \"kubernetes.io/projected/5da50628-e5d7-42e6-a221-f8503afc6885-kube-api-access-q5dng\") pod \"barbican-db-create-m88q9\" (UID: \"5da50628-e5d7-42e6-a221-f8503afc6885\") " pod="openstack/barbican-db-create-m88q9" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.477356 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9gk2n"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.478830 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9gk2n" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.523513 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9gk2n"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.553124 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-nxr4c"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.555024 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.555218 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5dng\" (UniqueName: \"kubernetes.io/projected/5da50628-e5d7-42e6-a221-f8503afc6885-kube-api-access-q5dng\") pod \"barbican-db-create-m88q9\" (UID: \"5da50628-e5d7-42e6-a221-f8503afc6885\") " pod="openstack/barbican-db-create-m88q9" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.572993 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vnkdb" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.573338 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.573470 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.573610 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.575358 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-nxr4c"] Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.631410 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5dng\" (UniqueName: \"kubernetes.io/projected/5da50628-e5d7-42e6-a221-f8503afc6885-kube-api-access-q5dng\") pod \"barbican-db-create-m88q9\" (UID: \"5da50628-e5d7-42e6-a221-f8503afc6885\") " pod="openstack/barbican-db-create-m88q9" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.660776 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-config-data\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.660840 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqntp\" (UniqueName: \"kubernetes.io/projected/5e35e950-34b7-4736-b669-6d3c52d0c797-kube-api-access-gqntp\") pod \"neutron-db-create-9gk2n\" (UID: \"5e35e950-34b7-4736-b669-6d3c52d0c797\") " pod="openstack/neutron-db-create-9gk2n" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.661017 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-combined-ca-bundle\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.661065 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6txp\" (UniqueName: \"kubernetes.io/projected/d3469c1a-9cfa-4d09-9672-40de98632e0e-kube-api-access-h6txp\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.682711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4d9n4" event={"ID":"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b","Type":"ContainerStarted","Data":"d4f5e3d31e35f437fd7a5ea9259d85f3d7be6c3703cdd068696ab1a4084a0e5e"} Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.685531 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2kvml-config-4h5lr" event={"ID":"09c40c7c-8032-4c68-b038-b427b089a157","Type":"ContainerStarted","Data":"17140782470f4f696b5460d53f84360e34c29ca973b7bc92ab7b5e523d164fc6"} Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.700476 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2kvml" podUID="cbefa245-5ad5-4303-b166-59574e8a0c65" containerName="ovn-controller" probeResult="failure" output=< Sep 30 14:38:47 crc kubenswrapper[4799]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Sep 30 14:38:47 crc kubenswrapper[4799]: > Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.714047 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4d9n4" podStartSLOduration=2.903906247 podStartE2EDuration="18.714018s" podCreationTimestamp="2025-09-30 14:38:29 +0000 UTC" firstStartedPulling="2025-09-30 14:38:30.593315824 +0000 UTC m=+1132.676916251" lastFinishedPulling="2025-09-30 14:38:46.403427577 +0000 UTC m=+1148.487028004" observedRunningTime="2025-09-30 14:38:47.708856169 +0000 UTC m=+1149.792456596" watchObservedRunningTime="2025-09-30 14:38:47.714018 +0000 UTC m=+1149.797618427" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.741178 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m88q9" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.762661 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-config-data\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.762715 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqntp\" (UniqueName: \"kubernetes.io/projected/5e35e950-34b7-4736-b669-6d3c52d0c797-kube-api-access-gqntp\") pod \"neutron-db-create-9gk2n\" (UID: \"5e35e950-34b7-4736-b669-6d3c52d0c797\") " pod="openstack/neutron-db-create-9gk2n" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.762790 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-combined-ca-bundle\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.762825 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6txp\" (UniqueName: \"kubernetes.io/projected/d3469c1a-9cfa-4d09-9672-40de98632e0e-kube-api-access-h6txp\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.787582 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-config-data\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.788340 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-combined-ca-bundle\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.791769 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6txp\" (UniqueName: \"kubernetes.io/projected/d3469c1a-9cfa-4d09-9672-40de98632e0e-kube-api-access-h6txp\") pod \"keystone-db-sync-nxr4c\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.809641 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqntp\" (UniqueName: \"kubernetes.io/projected/5e35e950-34b7-4736-b669-6d3c52d0c797-kube-api-access-gqntp\") pod \"neutron-db-create-9gk2n\" (UID: \"5e35e950-34b7-4736-b669-6d3c52d0c797\") " pod="openstack/neutron-db-create-9gk2n" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.963295 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9gk2n" Sep 30 14:38:47 crc kubenswrapper[4799]: I0930 14:38:47.982579 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:38:48 crc kubenswrapper[4799]: I0930 14:38:48.232039 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-7gkqb"] Sep 30 14:38:48 crc kubenswrapper[4799]: W0930 14:38:48.487172 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bb77af6_6c03_44f8_bec8_d168840a0c1e.slice/crio-c43120c95ecfc31cfb7befaa244f52f78abf3292b37a81f43f459bd1520e7d62 WatchSource:0}: Error finding container c43120c95ecfc31cfb7befaa244f52f78abf3292b37a81f43f459bd1520e7d62: Status 404 returned error can't find the container with id c43120c95ecfc31cfb7befaa244f52f78abf3292b37a81f43f459bd1520e7d62 Sep 30 14:38:48 crc kubenswrapper[4799]: I0930 14:38:48.746867 4799 generic.go:334] "Generic (PLEG): container finished" podID="09c40c7c-8032-4c68-b038-b427b089a157" containerID="5090e4ee5ee0dc23dd5819cb106268ef2e915599e7e70f6901c27df62a123e6f" exitCode=0 Sep 30 14:38:48 crc kubenswrapper[4799]: I0930 14:38:48.747247 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2kvml-config-4h5lr" event={"ID":"09c40c7c-8032-4c68-b038-b427b089a157","Type":"ContainerDied","Data":"5090e4ee5ee0dc23dd5819cb106268ef2e915599e7e70f6901c27df62a123e6f"} Sep 30 14:38:48 crc kubenswrapper[4799]: I0930 14:38:48.784900 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7gkqb" event={"ID":"8bb77af6-6c03-44f8-bec8-d168840a0c1e","Type":"ContainerStarted","Data":"c43120c95ecfc31cfb7befaa244f52f78abf3292b37a81f43f459bd1520e7d62"} Sep 30 14:38:49 crc kubenswrapper[4799]: I0930 14:38:49.199240 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m88q9"] Sep 30 14:38:49 crc kubenswrapper[4799]: I0930 14:38:49.317430 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9gk2n"] Sep 30 14:38:49 crc kubenswrapper[4799]: I0930 14:38:49.407050 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-nxr4c"] Sep 30 14:38:49 crc kubenswrapper[4799]: I0930 14:38:49.796528 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nxr4c" event={"ID":"d3469c1a-9cfa-4d09-9672-40de98632e0e","Type":"ContainerStarted","Data":"e111c33be1b0eb3f300da8866a85a988f53eec351bcdd8b68fa57f801a5d91fe"} Sep 30 14:38:49 crc kubenswrapper[4799]: I0930 14:38:49.798920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9gk2n" event={"ID":"5e35e950-34b7-4736-b669-6d3c52d0c797","Type":"ContainerStarted","Data":"a29f6bc01fa1cb1694d0690429c8963361def68656290f0acbfb8caacdfef254"} Sep 30 14:38:49 crc kubenswrapper[4799]: I0930 14:38:49.800554 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7gkqb" event={"ID":"8bb77af6-6c03-44f8-bec8-d168840a0c1e","Type":"ContainerStarted","Data":"cee3108a6408931221e66323254f6d4dc416ae87a8cf6f46f06f34655a58df6f"} Sep 30 14:38:49 crc kubenswrapper[4799]: I0930 14:38:49.807303 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m88q9" event={"ID":"5da50628-e5d7-42e6-a221-f8503afc6885","Type":"ContainerStarted","Data":"5c0c705a902af70b8d36cf4ca2f47fc5787cb5c3a7e1aaf5c111879e68572c10"} Sep 30 14:38:49 crc kubenswrapper[4799]: I0930 14:38:49.827213 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-7gkqb" podStartSLOduration=2.827188977 podStartE2EDuration="2.827188977s" podCreationTimestamp="2025-09-30 14:38:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:38:49.822620864 +0000 UTC m=+1151.906221301" watchObservedRunningTime="2025-09-30 14:38:49.827188977 +0000 UTC m=+1151.910789394" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.081685 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.216075 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run\") pod \"09c40c7c-8032-4c68-b038-b427b089a157\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.216127 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-additional-scripts\") pod \"09c40c7c-8032-4c68-b038-b427b089a157\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.216153 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run-ovn\") pod \"09c40c7c-8032-4c68-b038-b427b089a157\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.216238 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-log-ovn\") pod \"09c40c7c-8032-4c68-b038-b427b089a157\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.216271 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-scripts\") pod \"09c40c7c-8032-4c68-b038-b427b089a157\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.216298 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stb26\" (UniqueName: \"kubernetes.io/projected/09c40c7c-8032-4c68-b038-b427b089a157-kube-api-access-stb26\") pod \"09c40c7c-8032-4c68-b038-b427b089a157\" (UID: \"09c40c7c-8032-4c68-b038-b427b089a157\") " Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.218053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "09c40c7c-8032-4c68-b038-b427b089a157" (UID: "09c40c7c-8032-4c68-b038-b427b089a157"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.218104 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run" (OuterVolumeSpecName: "var-run") pod "09c40c7c-8032-4c68-b038-b427b089a157" (UID: "09c40c7c-8032-4c68-b038-b427b089a157"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.218618 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "09c40c7c-8032-4c68-b038-b427b089a157" (UID: "09c40c7c-8032-4c68-b038-b427b089a157"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.218690 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "09c40c7c-8032-4c68-b038-b427b089a157" (UID: "09c40c7c-8032-4c68-b038-b427b089a157"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.219308 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-scripts" (OuterVolumeSpecName: "scripts") pod "09c40c7c-8032-4c68-b038-b427b089a157" (UID: "09c40c7c-8032-4c68-b038-b427b089a157"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.231865 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09c40c7c-8032-4c68-b038-b427b089a157-kube-api-access-stb26" (OuterVolumeSpecName: "kube-api-access-stb26") pod "09c40c7c-8032-4c68-b038-b427b089a157" (UID: "09c40c7c-8032-4c68-b038-b427b089a157"). InnerVolumeSpecName "kube-api-access-stb26". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.317949 4799 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.317993 4799 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-additional-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.318010 4799 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.318021 4799 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/09c40c7c-8032-4c68-b038-b427b089a157-var-log-ovn\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.318033 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09c40c7c-8032-4c68-b038-b427b089a157-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.318043 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stb26\" (UniqueName: \"kubernetes.io/projected/09c40c7c-8032-4c68-b038-b427b089a157-kube-api-access-stb26\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.839727 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2kvml-config-4h5lr" event={"ID":"09c40c7c-8032-4c68-b038-b427b089a157","Type":"ContainerDied","Data":"17140782470f4f696b5460d53f84360e34c29ca973b7bc92ab7b5e523d164fc6"} Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.840100 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17140782470f4f696b5460d53f84360e34c29ca973b7bc92ab7b5e523d164fc6" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.840048 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2kvml-config-4h5lr" Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.847945 4799 generic.go:334] "Generic (PLEG): container finished" podID="5e35e950-34b7-4736-b669-6d3c52d0c797" containerID="f2cebab7abd7a7cc43f8ebbf8af08ab2a140cb6467343873cc3c86e255405695" exitCode=0 Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.849230 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9gk2n" event={"ID":"5e35e950-34b7-4736-b669-6d3c52d0c797","Type":"ContainerDied","Data":"f2cebab7abd7a7cc43f8ebbf8af08ab2a140cb6467343873cc3c86e255405695"} Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.855804 4799 generic.go:334] "Generic (PLEG): container finished" podID="8bb77af6-6c03-44f8-bec8-d168840a0c1e" containerID="cee3108a6408931221e66323254f6d4dc416ae87a8cf6f46f06f34655a58df6f" exitCode=0 Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.855955 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7gkqb" event={"ID":"8bb77af6-6c03-44f8-bec8-d168840a0c1e","Type":"ContainerDied","Data":"cee3108a6408931221e66323254f6d4dc416ae87a8cf6f46f06f34655a58df6f"} Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.859940 4799 generic.go:334] "Generic (PLEG): container finished" podID="5da50628-e5d7-42e6-a221-f8503afc6885" containerID="3ba1f87126afcc820da7f7ca3ba6d11ef72a1bc8d3a4d7c1f99d838c77ee19b2" exitCode=0 Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.860022 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m88q9" event={"ID":"5da50628-e5d7-42e6-a221-f8503afc6885","Type":"ContainerDied","Data":"3ba1f87126afcc820da7f7ca3ba6d11ef72a1bc8d3a4d7c1f99d838c77ee19b2"} Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.875763 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"0e3078ffb48ddb95cc98cac4e76dba3be008409b82f3e6b1a82031a9c6b9bef8"} Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.875830 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"043c66ef5177b962734a9e3640fc4b9ad1cc0dd03d6273d9bbcc26693418da81"} Sep 30 14:38:50 crc kubenswrapper[4799]: I0930 14:38:50.875841 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"0856ea1a6db46b333ee197a3bd95157e3f448192320bb18b3584c01009761836"} Sep 30 14:38:51 crc kubenswrapper[4799]: I0930 14:38:51.214844 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2kvml-config-4h5lr"] Sep 30 14:38:51 crc kubenswrapper[4799]: I0930 14:38:51.222118 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2kvml-config-4h5lr"] Sep 30 14:38:51 crc kubenswrapper[4799]: I0930 14:38:51.906925 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"79b368c798095e1a1aa98b5a561f0c88f29bee0f0d175ff68143702e011530fa"} Sep 30 14:38:51 crc kubenswrapper[4799]: I0930 14:38:51.907393 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"9897ad316832c0a379f2df194d402a97d48d2d371b8a9e926ae1e3969af95b3c"} Sep 30 14:38:51 crc kubenswrapper[4799]: I0930 14:38:51.907415 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"bf95b530f387fd3ce6795af0e0bd7fd07622ff249cb72fc63e7e91715db5da58"} Sep 30 14:38:51 crc kubenswrapper[4799]: I0930 14:38:51.907430 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1500572d-64d4-483a-a525-bc0530fe4cbb","Type":"ContainerStarted","Data":"2a8583efca2e245c157cb1b6b620a7a383ed2c2ec7c328ee3bf53ca513844c7f"} Sep 30 14:38:51 crc kubenswrapper[4799]: I0930 14:38:51.954081 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.528770411 podStartE2EDuration="40.954056203s" podCreationTimestamp="2025-09-30 14:38:11 +0000 UTC" firstStartedPulling="2025-09-30 14:38:29.11575966 +0000 UTC m=+1131.199360087" lastFinishedPulling="2025-09-30 14:38:49.541045452 +0000 UTC m=+1151.624645879" observedRunningTime="2025-09-30 14:38:51.95257411 +0000 UTC m=+1154.036174557" watchObservedRunningTime="2025-09-30 14:38:51.954056203 +0000 UTC m=+1154.037656630" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.325414 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6dhwn"] Sep 30 14:38:52 crc kubenswrapper[4799]: E0930 14:38:52.325918 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09c40c7c-8032-4c68-b038-b427b089a157" containerName="ovn-config" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.325935 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="09c40c7c-8032-4c68-b038-b427b089a157" containerName="ovn-config" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.326158 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="09c40c7c-8032-4c68-b038-b427b089a157" containerName="ovn-config" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.327354 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.334383 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.349354 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6dhwn"] Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.376868 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.377270 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.377319 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.377346 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.377711 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzj9w\" (UniqueName: \"kubernetes.io/projected/a98d7efa-190b-4499-b6ba-85aadf66ecae-kube-api-access-wzj9w\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.377788 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.487563 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.487969 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.487998 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.488024 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.488076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzj9w\" (UniqueName: \"kubernetes.io/projected/a98d7efa-190b-4499-b6ba-85aadf66ecae-kube-api-access-wzj9w\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.488098 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.488625 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.494432 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.495127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.498304 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.499360 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.533773 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzj9w\" (UniqueName: \"kubernetes.io/projected/a98d7efa-190b-4499-b6ba-85aadf66ecae-kube-api-access-wzj9w\") pod \"dnsmasq-dns-77585f5f8c-6dhwn\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.573701 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09c40c7c-8032-4c68-b038-b427b089a157" path="/var/lib/kubelet/pods/09c40c7c-8032-4c68-b038-b427b089a157/volumes" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.738547 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.743392 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m88q9" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.776390 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9gk2n" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.807779 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7gkqb" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.809285 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqntp\" (UniqueName: \"kubernetes.io/projected/5e35e950-34b7-4736-b669-6d3c52d0c797-kube-api-access-gqntp\") pod \"5e35e950-34b7-4736-b669-6d3c52d0c797\" (UID: \"5e35e950-34b7-4736-b669-6d3c52d0c797\") " Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.809417 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5dng\" (UniqueName: \"kubernetes.io/projected/5da50628-e5d7-42e6-a221-f8503afc6885-kube-api-access-q5dng\") pod \"5da50628-e5d7-42e6-a221-f8503afc6885\" (UID: \"5da50628-e5d7-42e6-a221-f8503afc6885\") " Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.838508 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da50628-e5d7-42e6-a221-f8503afc6885-kube-api-access-q5dng" (OuterVolumeSpecName: "kube-api-access-q5dng") pod "5da50628-e5d7-42e6-a221-f8503afc6885" (UID: "5da50628-e5d7-42e6-a221-f8503afc6885"). InnerVolumeSpecName "kube-api-access-q5dng". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.862227 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e35e950-34b7-4736-b669-6d3c52d0c797-kube-api-access-gqntp" (OuterVolumeSpecName: "kube-api-access-gqntp") pod "5e35e950-34b7-4736-b669-6d3c52d0c797" (UID: "5e35e950-34b7-4736-b669-6d3c52d0c797"). InnerVolumeSpecName "kube-api-access-gqntp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.872434 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-2kvml" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.912065 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-452wh\" (UniqueName: \"kubernetes.io/projected/8bb77af6-6c03-44f8-bec8-d168840a0c1e-kube-api-access-452wh\") pod \"8bb77af6-6c03-44f8-bec8-d168840a0c1e\" (UID: \"8bb77af6-6c03-44f8-bec8-d168840a0c1e\") " Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.912894 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqntp\" (UniqueName: \"kubernetes.io/projected/5e35e950-34b7-4736-b669-6d3c52d0c797-kube-api-access-gqntp\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.912919 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5dng\" (UniqueName: \"kubernetes.io/projected/5da50628-e5d7-42e6-a221-f8503afc6885-kube-api-access-q5dng\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.924868 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bb77af6-6c03-44f8-bec8-d168840a0c1e-kube-api-access-452wh" (OuterVolumeSpecName: "kube-api-access-452wh") pod "8bb77af6-6c03-44f8-bec8-d168840a0c1e" (UID: "8bb77af6-6c03-44f8-bec8-d168840a0c1e"). InnerVolumeSpecName "kube-api-access-452wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.933637 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9gk2n" event={"ID":"5e35e950-34b7-4736-b669-6d3c52d0c797","Type":"ContainerDied","Data":"a29f6bc01fa1cb1694d0690429c8963361def68656290f0acbfb8caacdfef254"} Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.933699 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a29f6bc01fa1cb1694d0690429c8963361def68656290f0acbfb8caacdfef254" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.933772 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9gk2n" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.944001 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-7gkqb" event={"ID":"8bb77af6-6c03-44f8-bec8-d168840a0c1e","Type":"ContainerDied","Data":"c43120c95ecfc31cfb7befaa244f52f78abf3292b37a81f43f459bd1520e7d62"} Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.944107 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c43120c95ecfc31cfb7befaa244f52f78abf3292b37a81f43f459bd1520e7d62" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.944218 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-7gkqb" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.953111 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m88q9" Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.953216 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m88q9" event={"ID":"5da50628-e5d7-42e6-a221-f8503afc6885","Type":"ContainerDied","Data":"5c0c705a902af70b8d36cf4ca2f47fc5787cb5c3a7e1aaf5c111879e68572c10"} Sep 30 14:38:52 crc kubenswrapper[4799]: I0930 14:38:52.953266 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c0c705a902af70b8d36cf4ca2f47fc5787cb5c3a7e1aaf5c111879e68572c10" Sep 30 14:38:53 crc kubenswrapper[4799]: I0930 14:38:53.030443 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-452wh\" (UniqueName: \"kubernetes.io/projected/8bb77af6-6c03-44f8-bec8-d168840a0c1e-kube-api-access-452wh\") on node \"crc\" DevicePath \"\"" Sep 30 14:38:53 crc kubenswrapper[4799]: I0930 14:38:53.464500 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6dhwn"] Sep 30 14:38:53 crc kubenswrapper[4799]: I0930 14:38:53.967776 4799 generic.go:334] "Generic (PLEG): container finished" podID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerID="febc20fa98082e3d1a188cbe33e9267d0a6f2de59c8f0288fa1227b75ebb2f05" exitCode=0 Sep 30 14:38:53 crc kubenswrapper[4799]: I0930 14:38:53.967933 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" event={"ID":"a98d7efa-190b-4499-b6ba-85aadf66ecae","Type":"ContainerDied","Data":"febc20fa98082e3d1a188cbe33e9267d0a6f2de59c8f0288fa1227b75ebb2f05"} Sep 30 14:38:53 crc kubenswrapper[4799]: I0930 14:38:53.968157 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" event={"ID":"a98d7efa-190b-4499-b6ba-85aadf66ecae","Type":"ContainerStarted","Data":"e03c857718c12876b33b8d6f3bb8d857c024b688d14d81495c6dce6fde665c94"} Sep 30 14:39:01 crc kubenswrapper[4799]: I0930 14:39:01.050186 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" event={"ID":"a98d7efa-190b-4499-b6ba-85aadf66ecae","Type":"ContainerStarted","Data":"a8e149f9114058145cb9cc5f486f2cff9cd522f651cd10b8d2b4cd814d21d9a1"} Sep 30 14:39:01 crc kubenswrapper[4799]: I0930 14:39:01.051791 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:39:01 crc kubenswrapper[4799]: I0930 14:39:01.080104 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" podStartSLOduration=9.080071393 podStartE2EDuration="9.080071393s" podCreationTimestamp="2025-09-30 14:38:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:01.075805148 +0000 UTC m=+1163.159405585" watchObservedRunningTime="2025-09-30 14:39:01.080071393 +0000 UTC m=+1163.163671820" Sep 30 14:39:02 crc kubenswrapper[4799]: I0930 14:39:02.060079 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nxr4c" event={"ID":"d3469c1a-9cfa-4d09-9672-40de98632e0e","Type":"ContainerStarted","Data":"a0efc0d12665083a1638a4b9241694233639c17b85a2484725aa6a18f922f2b7"} Sep 30 14:39:06 crc kubenswrapper[4799]: I0930 14:39:06.096148 4799 generic.go:334] "Generic (PLEG): container finished" podID="bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" containerID="d4f5e3d31e35f437fd7a5ea9259d85f3d7be6c3703cdd068696ab1a4084a0e5e" exitCode=0 Sep 30 14:39:06 crc kubenswrapper[4799]: I0930 14:39:06.096232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4d9n4" event={"ID":"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b","Type":"ContainerDied","Data":"d4f5e3d31e35f437fd7a5ea9259d85f3d7be6c3703cdd068696ab1a4084a0e5e"} Sep 30 14:39:06 crc kubenswrapper[4799]: I0930 14:39:06.116758 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-nxr4c" podStartSLOduration=7.733543788 podStartE2EDuration="19.1166262s" podCreationTimestamp="2025-09-30 14:38:47 +0000 UTC" firstStartedPulling="2025-09-30 14:38:49.518020899 +0000 UTC m=+1151.601621326" lastFinishedPulling="2025-09-30 14:39:00.901103321 +0000 UTC m=+1162.984703738" observedRunningTime="2025-09-30 14:39:02.082122436 +0000 UTC m=+1164.165722873" watchObservedRunningTime="2025-09-30 14:39:06.1166262 +0000 UTC m=+1168.200226637" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.229055 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-4c10-account-create-r5fpd"] Sep 30 14:39:07 crc kubenswrapper[4799]: E0930 14:39:07.229453 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da50628-e5d7-42e6-a221-f8503afc6885" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.229469 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da50628-e5d7-42e6-a221-f8503afc6885" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: E0930 14:39:07.229486 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb77af6-6c03-44f8-bec8-d168840a0c1e" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.229492 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb77af6-6c03-44f8-bec8-d168840a0c1e" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: E0930 14:39:07.229509 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e35e950-34b7-4736-b669-6d3c52d0c797" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.229515 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e35e950-34b7-4736-b669-6d3c52d0c797" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.229700 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da50628-e5d7-42e6-a221-f8503afc6885" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.229715 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e35e950-34b7-4736-b669-6d3c52d0c797" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.229729 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb77af6-6c03-44f8-bec8-d168840a0c1e" containerName="mariadb-database-create" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.230389 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4c10-account-create-r5fpd" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.252894 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4c10-account-create-r5fpd"] Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.253901 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.384221 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsplw\" (UniqueName: \"kubernetes.io/projected/a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499-kube-api-access-fsplw\") pod \"cinder-4c10-account-create-r5fpd\" (UID: \"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499\") " pod="openstack/cinder-4c10-account-create-r5fpd" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.451664 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6f25-account-create-hcvsn"] Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.456390 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6f25-account-create-hcvsn" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.463300 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.465279 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6f25-account-create-hcvsn"] Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.490395 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsplw\" (UniqueName: \"kubernetes.io/projected/a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499-kube-api-access-fsplw\") pod \"cinder-4c10-account-create-r5fpd\" (UID: \"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499\") " pod="openstack/cinder-4c10-account-create-r5fpd" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.554776 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsplw\" (UniqueName: \"kubernetes.io/projected/a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499-kube-api-access-fsplw\") pod \"cinder-4c10-account-create-r5fpd\" (UID: \"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499\") " pod="openstack/cinder-4c10-account-create-r5fpd" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.560260 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4c10-account-create-r5fpd" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.599288 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwd89\" (UniqueName: \"kubernetes.io/projected/acec8647-568b-474c-86a6-590726b7b86e-kube-api-access-gwd89\") pod \"barbican-6f25-account-create-hcvsn\" (UID: \"acec8647-568b-474c-86a6-590726b7b86e\") " pod="openstack/barbican-6f25-account-create-hcvsn" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.643772 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ac6f-account-create-cfpcr"] Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.645334 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ac6f-account-create-cfpcr" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.648280 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.679895 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ac6f-account-create-cfpcr"] Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.703682 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwd89\" (UniqueName: \"kubernetes.io/projected/acec8647-568b-474c-86a6-590726b7b86e-kube-api-access-gwd89\") pod \"barbican-6f25-account-create-hcvsn\" (UID: \"acec8647-568b-474c-86a6-590726b7b86e\") " pod="openstack/barbican-6f25-account-create-hcvsn" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.703785 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnndf\" (UniqueName: \"kubernetes.io/projected/ea31ca1c-9064-41a9-a55e-24d09ee891f3-kube-api-access-xnndf\") pod \"neutron-ac6f-account-create-cfpcr\" (UID: \"ea31ca1c-9064-41a9-a55e-24d09ee891f3\") " pod="openstack/neutron-ac6f-account-create-cfpcr" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.748874 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.756033 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4d9n4" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.796750 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwd89\" (UniqueName: \"kubernetes.io/projected/acec8647-568b-474c-86a6-590726b7b86e-kube-api-access-gwd89\") pod \"barbican-6f25-account-create-hcvsn\" (UID: \"acec8647-568b-474c-86a6-590726b7b86e\") " pod="openstack/barbican-6f25-account-create-hcvsn" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.821334 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-db-sync-config-data\") pod \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.821817 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-config-data\") pod \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.821927 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-combined-ca-bundle\") pod \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.821995 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4kwk\" (UniqueName: \"kubernetes.io/projected/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-kube-api-access-n4kwk\") pod \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\" (UID: \"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b\") " Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.822316 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnndf\" (UniqueName: \"kubernetes.io/projected/ea31ca1c-9064-41a9-a55e-24d09ee891f3-kube-api-access-xnndf\") pod \"neutron-ac6f-account-create-cfpcr\" (UID: \"ea31ca1c-9064-41a9-a55e-24d09ee891f3\") " pod="openstack/neutron-ac6f-account-create-cfpcr" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.853550 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" (UID: "bfbbeb1f-d64a-46ac-a80d-3541ee813f6b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.869688 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-kube-api-access-n4kwk" (OuterVolumeSpecName: "kube-api-access-n4kwk") pod "bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" (UID: "bfbbeb1f-d64a-46ac-a80d-3541ee813f6b"). InnerVolumeSpecName "kube-api-access-n4kwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.882688 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnndf\" (UniqueName: \"kubernetes.io/projected/ea31ca1c-9064-41a9-a55e-24d09ee891f3-kube-api-access-xnndf\") pod \"neutron-ac6f-account-create-cfpcr\" (UID: \"ea31ca1c-9064-41a9-a55e-24d09ee891f3\") " pod="openstack/neutron-ac6f-account-create-cfpcr" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.919895 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kbqg4"] Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.920192 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-kbqg4" podUID="86f6895b-e518-4d47-8b9e-ad1143dbc433" containerName="dnsmasq-dns" containerID="cri-o://0230a08df166e7eec4cbc204bd89968f918187177b7c0d95489950ceb9f902f8" gracePeriod=10 Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.925707 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.925752 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4kwk\" (UniqueName: \"kubernetes.io/projected/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-kube-api-access-n4kwk\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.945620 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" (UID: "bfbbeb1f-d64a-46ac-a80d-3541ee813f6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:07 crc kubenswrapper[4799]: I0930 14:39:07.949113 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-config-data" (OuterVolumeSpecName: "config-data") pod "bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" (UID: "bfbbeb1f-d64a-46ac-a80d-3541ee813f6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.027923 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.028005 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.053176 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ac6f-account-create-cfpcr" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.091210 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6f25-account-create-hcvsn" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.116985 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4d9n4" event={"ID":"bfbbeb1f-d64a-46ac-a80d-3541ee813f6b","Type":"ContainerDied","Data":"3313cf6d4f81f4ac8c2ad5faddff4fb04ba86817410812a0c510e52b218de228"} Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.117030 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3313cf6d4f81f4ac8c2ad5faddff4fb04ba86817410812a0c510e52b218de228" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.117093 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4d9n4" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.443498 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4c10-account-create-r5fpd"] Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.667106 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ac6f-account-create-cfpcr"] Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.741837 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-s4n2b"] Sep 30 14:39:08 crc kubenswrapper[4799]: E0930 14:39:08.742823 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" containerName="glance-db-sync" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.742846 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" containerName="glance-db-sync" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.743064 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" containerName="glance-db-sync" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.756372 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.785257 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-s4n2b"] Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.902325 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.902405 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-config\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.902439 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqzdz\" (UniqueName: \"kubernetes.io/projected/bdb12c35-9f78-4995-b550-573981289693-kube-api-access-zqzdz\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.902477 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.902523 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.902545 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:08 crc kubenswrapper[4799]: I0930 14:39:08.914988 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6f25-account-create-hcvsn"] Sep 30 14:39:08 crc kubenswrapper[4799]: W0930 14:39:08.959504 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacec8647_568b_474c_86a6_590726b7b86e.slice/crio-56654f47ea838c3b10ca83e62a0b8d6a46e4119cdd1e1ad14717d2f2c287afbe WatchSource:0}: Error finding container 56654f47ea838c3b10ca83e62a0b8d6a46e4119cdd1e1ad14717d2f2c287afbe: Status 404 returned error can't find the container with id 56654f47ea838c3b10ca83e62a0b8d6a46e4119cdd1e1ad14717d2f2c287afbe Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.005272 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-config\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.005335 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqzdz\" (UniqueName: \"kubernetes.io/projected/bdb12c35-9f78-4995-b550-573981289693-kube-api-access-zqzdz\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.005381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.005429 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.005482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.005513 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.006513 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.007166 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-config\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.008052 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.008563 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.009506 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.033702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqzdz\" (UniqueName: \"kubernetes.io/projected/bdb12c35-9f78-4995-b550-573981289693-kube-api-access-zqzdz\") pod \"dnsmasq-dns-7ff5475cc9-s4n2b\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.115394 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.156454 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ac6f-account-create-cfpcr" event={"ID":"ea31ca1c-9064-41a9-a55e-24d09ee891f3","Type":"ContainerStarted","Data":"413c359b00855aeb89ae340b26ae9452dce56269bd92cb960e5dffc6735e24d2"} Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.156900 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ac6f-account-create-cfpcr" event={"ID":"ea31ca1c-9064-41a9-a55e-24d09ee891f3","Type":"ContainerStarted","Data":"aec1e92892ccfac7c397a352343e98bdd6f38314aa7d86cddd5db55dad37088e"} Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.162932 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6f25-account-create-hcvsn" event={"ID":"acec8647-568b-474c-86a6-590726b7b86e","Type":"ContainerStarted","Data":"56654f47ea838c3b10ca83e62a0b8d6a46e4119cdd1e1ad14717d2f2c287afbe"} Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.227245 4799 generic.go:334] "Generic (PLEG): container finished" podID="86f6895b-e518-4d47-8b9e-ad1143dbc433" containerID="0230a08df166e7eec4cbc204bd89968f918187177b7c0d95489950ceb9f902f8" exitCode=0 Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.227329 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kbqg4" event={"ID":"86f6895b-e518-4d47-8b9e-ad1143dbc433","Type":"ContainerDied","Data":"0230a08df166e7eec4cbc204bd89968f918187177b7c0d95489950ceb9f902f8"} Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.239576 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4c10-account-create-r5fpd" event={"ID":"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499","Type":"ContainerStarted","Data":"a7cb6d772c2f7b01603f66796135de6d05559ae3bef5e92adda80722fa0bd601"} Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.239623 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4c10-account-create-r5fpd" event={"ID":"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499","Type":"ContainerStarted","Data":"738f36f55e46138e84aae57da807ec6a18b119eb6531a88e415be1d2aab894fc"} Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.261215 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-4c10-account-create-r5fpd" podStartSLOduration=2.261190758 podStartE2EDuration="2.261190758s" podCreationTimestamp="2025-09-30 14:39:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:09.259468687 +0000 UTC m=+1171.343069114" watchObservedRunningTime="2025-09-30 14:39:09.261190758 +0000 UTC m=+1171.344791185" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.267559 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-ac6f-account-create-cfpcr" podStartSLOduration=2.267524603 podStartE2EDuration="2.267524603s" podCreationTimestamp="2025-09-30 14:39:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:09.187141193 +0000 UTC m=+1171.270741620" watchObservedRunningTime="2025-09-30 14:39:09.267524603 +0000 UTC m=+1171.351125030" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.302896 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.348504 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-dns-svc\") pod \"86f6895b-e518-4d47-8b9e-ad1143dbc433\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.348665 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvnxw\" (UniqueName: \"kubernetes.io/projected/86f6895b-e518-4d47-8b9e-ad1143dbc433-kube-api-access-pvnxw\") pod \"86f6895b-e518-4d47-8b9e-ad1143dbc433\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.348772 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-nb\") pod \"86f6895b-e518-4d47-8b9e-ad1143dbc433\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.348845 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-config\") pod \"86f6895b-e518-4d47-8b9e-ad1143dbc433\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.348880 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-sb\") pod \"86f6895b-e518-4d47-8b9e-ad1143dbc433\" (UID: \"86f6895b-e518-4d47-8b9e-ad1143dbc433\") " Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.377988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86f6895b-e518-4d47-8b9e-ad1143dbc433-kube-api-access-pvnxw" (OuterVolumeSpecName: "kube-api-access-pvnxw") pod "86f6895b-e518-4d47-8b9e-ad1143dbc433" (UID: "86f6895b-e518-4d47-8b9e-ad1143dbc433"). InnerVolumeSpecName "kube-api-access-pvnxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.437495 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-config" (OuterVolumeSpecName: "config") pod "86f6895b-e518-4d47-8b9e-ad1143dbc433" (UID: "86f6895b-e518-4d47-8b9e-ad1143dbc433"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.451028 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvnxw\" (UniqueName: \"kubernetes.io/projected/86f6895b-e518-4d47-8b9e-ad1143dbc433-kube-api-access-pvnxw\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.452114 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.491309 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "86f6895b-e518-4d47-8b9e-ad1143dbc433" (UID: "86f6895b-e518-4d47-8b9e-ad1143dbc433"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.497842 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "86f6895b-e518-4d47-8b9e-ad1143dbc433" (UID: "86f6895b-e518-4d47-8b9e-ad1143dbc433"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.530204 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "86f6895b-e518-4d47-8b9e-ad1143dbc433" (UID: "86f6895b-e518-4d47-8b9e-ad1143dbc433"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.558124 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.558168 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.558183 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86f6895b-e518-4d47-8b9e-ad1143dbc433-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:09 crc kubenswrapper[4799]: I0930 14:39:09.868044 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-s4n2b"] Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.270064 4799 generic.go:334] "Generic (PLEG): container finished" podID="acec8647-568b-474c-86a6-590726b7b86e" containerID="d3f69fa4bdfb6650ca83eb0148fd6757e7c8a958c66b2c3af14ed14c089a5189" exitCode=0 Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.270485 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6f25-account-create-hcvsn" event={"ID":"acec8647-568b-474c-86a6-590726b7b86e","Type":"ContainerDied","Data":"d3f69fa4bdfb6650ca83eb0148fd6757e7c8a958c66b2c3af14ed14c089a5189"} Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.291413 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kbqg4" event={"ID":"86f6895b-e518-4d47-8b9e-ad1143dbc433","Type":"ContainerDied","Data":"378a5510f7c553447365e6cc0f9f777870de06ed4a7498339fdaa58d321c039c"} Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.291836 4799 scope.go:117] "RemoveContainer" containerID="0230a08df166e7eec4cbc204bd89968f918187177b7c0d95489950ceb9f902f8" Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.292202 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kbqg4" Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.309084 4799 generic.go:334] "Generic (PLEG): container finished" podID="a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499" containerID="a7cb6d772c2f7b01603f66796135de6d05559ae3bef5e92adda80722fa0bd601" exitCode=0 Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.309222 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4c10-account-create-r5fpd" event={"ID":"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499","Type":"ContainerDied","Data":"a7cb6d772c2f7b01603f66796135de6d05559ae3bef5e92adda80722fa0bd601"} Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.315729 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" event={"ID":"bdb12c35-9f78-4995-b550-573981289693","Type":"ContainerStarted","Data":"0ce0147026b273b9297927ca0054d3adcd9b22e55cdd543329a1e06d5193e758"} Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.315988 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" event={"ID":"bdb12c35-9f78-4995-b550-573981289693","Type":"ContainerStarted","Data":"a24a8ad4d1dd96357d41fedc66a5211355fb104bf59b74a10fc8eeff7e1c443d"} Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.332945 4799 generic.go:334] "Generic (PLEG): container finished" podID="ea31ca1c-9064-41a9-a55e-24d09ee891f3" containerID="413c359b00855aeb89ae340b26ae9452dce56269bd92cb960e5dffc6735e24d2" exitCode=0 Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.333268 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ac6f-account-create-cfpcr" event={"ID":"ea31ca1c-9064-41a9-a55e-24d09ee891f3","Type":"ContainerDied","Data":"413c359b00855aeb89ae340b26ae9452dce56269bd92cb960e5dffc6735e24d2"} Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.337292 4799 scope.go:117] "RemoveContainer" containerID="df84b6ea18e8be02b814045e5d16038b565e6b0ba8ea61cadfc5e62f05c13ace" Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.378058 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kbqg4"] Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.389608 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kbqg4"] Sep 30 14:39:10 crc kubenswrapper[4799]: I0930 14:39:10.514316 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86f6895b-e518-4d47-8b9e-ad1143dbc433" path="/var/lib/kubelet/pods/86f6895b-e518-4d47-8b9e-ad1143dbc433/volumes" Sep 30 14:39:11 crc kubenswrapper[4799]: I0930 14:39:11.343524 4799 generic.go:334] "Generic (PLEG): container finished" podID="bdb12c35-9f78-4995-b550-573981289693" containerID="0ce0147026b273b9297927ca0054d3adcd9b22e55cdd543329a1e06d5193e758" exitCode=0 Sep 30 14:39:11 crc kubenswrapper[4799]: I0930 14:39:11.343599 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" event={"ID":"bdb12c35-9f78-4995-b550-573981289693","Type":"ContainerDied","Data":"0ce0147026b273b9297927ca0054d3adcd9b22e55cdd543329a1e06d5193e758"} Sep 30 14:39:11 crc kubenswrapper[4799]: I0930 14:39:11.899987 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6f25-account-create-hcvsn" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.018737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwd89\" (UniqueName: \"kubernetes.io/projected/acec8647-568b-474c-86a6-590726b7b86e-kube-api-access-gwd89\") pod \"acec8647-568b-474c-86a6-590726b7b86e\" (UID: \"acec8647-568b-474c-86a6-590726b7b86e\") " Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.028132 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acec8647-568b-474c-86a6-590726b7b86e-kube-api-access-gwd89" (OuterVolumeSpecName: "kube-api-access-gwd89") pod "acec8647-568b-474c-86a6-590726b7b86e" (UID: "acec8647-568b-474c-86a6-590726b7b86e"). InnerVolumeSpecName "kube-api-access-gwd89". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.120401 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwd89\" (UniqueName: \"kubernetes.io/projected/acec8647-568b-474c-86a6-590726b7b86e-kube-api-access-gwd89\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.122138 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4c10-account-create-r5fpd" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.139347 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ac6f-account-create-cfpcr" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.221686 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsplw\" (UniqueName: \"kubernetes.io/projected/a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499-kube-api-access-fsplw\") pod \"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499\" (UID: \"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499\") " Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.221972 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnndf\" (UniqueName: \"kubernetes.io/projected/ea31ca1c-9064-41a9-a55e-24d09ee891f3-kube-api-access-xnndf\") pod \"ea31ca1c-9064-41a9-a55e-24d09ee891f3\" (UID: \"ea31ca1c-9064-41a9-a55e-24d09ee891f3\") " Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.227691 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499-kube-api-access-fsplw" (OuterVolumeSpecName: "kube-api-access-fsplw") pod "a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499" (UID: "a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499"). InnerVolumeSpecName "kube-api-access-fsplw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.235138 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea31ca1c-9064-41a9-a55e-24d09ee891f3-kube-api-access-xnndf" (OuterVolumeSpecName: "kube-api-access-xnndf") pod "ea31ca1c-9064-41a9-a55e-24d09ee891f3" (UID: "ea31ca1c-9064-41a9-a55e-24d09ee891f3"). InnerVolumeSpecName "kube-api-access-xnndf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.324406 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnndf\" (UniqueName: \"kubernetes.io/projected/ea31ca1c-9064-41a9-a55e-24d09ee891f3-kube-api-access-xnndf\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.324838 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsplw\" (UniqueName: \"kubernetes.io/projected/a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499-kube-api-access-fsplw\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.357410 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ac6f-account-create-cfpcr" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.357349 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ac6f-account-create-cfpcr" event={"ID":"ea31ca1c-9064-41a9-a55e-24d09ee891f3","Type":"ContainerDied","Data":"aec1e92892ccfac7c397a352343e98bdd6f38314aa7d86cddd5db55dad37088e"} Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.357923 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aec1e92892ccfac7c397a352343e98bdd6f38314aa7d86cddd5db55dad37088e" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.359913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6f25-account-create-hcvsn" event={"ID":"acec8647-568b-474c-86a6-590726b7b86e","Type":"ContainerDied","Data":"56654f47ea838c3b10ca83e62a0b8d6a46e4119cdd1e1ad14717d2f2c287afbe"} Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.359957 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56654f47ea838c3b10ca83e62a0b8d6a46e4119cdd1e1ad14717d2f2c287afbe" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.359959 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6f25-account-create-hcvsn" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.362190 4799 generic.go:334] "Generic (PLEG): container finished" podID="d3469c1a-9cfa-4d09-9672-40de98632e0e" containerID="a0efc0d12665083a1638a4b9241694233639c17b85a2484725aa6a18f922f2b7" exitCode=0 Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.362254 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nxr4c" event={"ID":"d3469c1a-9cfa-4d09-9672-40de98632e0e","Type":"ContainerDied","Data":"a0efc0d12665083a1638a4b9241694233639c17b85a2484725aa6a18f922f2b7"} Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.365629 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" event={"ID":"bdb12c35-9f78-4995-b550-573981289693","Type":"ContainerStarted","Data":"1fdbd0951f319d68c563518077053a1e2f091c7f8100221c9286ff24c460b152"} Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.366796 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.370558 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4c10-account-create-r5fpd" event={"ID":"a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499","Type":"ContainerDied","Data":"738f36f55e46138e84aae57da807ec6a18b119eb6531a88e415be1d2aab894fc"} Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.370626 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="738f36f55e46138e84aae57da807ec6a18b119eb6531a88e415be1d2aab894fc" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.370689 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4c10-account-create-r5fpd" Sep 30 14:39:12 crc kubenswrapper[4799]: I0930 14:39:12.430917 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" podStartSLOduration=4.43089628 podStartE2EDuration="4.43089628s" podCreationTimestamp="2025-09-30 14:39:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:12.429036306 +0000 UTC m=+1174.512636753" watchObservedRunningTime="2025-09-30 14:39:12.43089628 +0000 UTC m=+1174.514496707" Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.746307 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.853610 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-combined-ca-bundle\") pod \"d3469c1a-9cfa-4d09-9672-40de98632e0e\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.853828 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-config-data\") pod \"d3469c1a-9cfa-4d09-9672-40de98632e0e\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.853888 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6txp\" (UniqueName: \"kubernetes.io/projected/d3469c1a-9cfa-4d09-9672-40de98632e0e-kube-api-access-h6txp\") pod \"d3469c1a-9cfa-4d09-9672-40de98632e0e\" (UID: \"d3469c1a-9cfa-4d09-9672-40de98632e0e\") " Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.858276 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3469c1a-9cfa-4d09-9672-40de98632e0e-kube-api-access-h6txp" (OuterVolumeSpecName: "kube-api-access-h6txp") pod "d3469c1a-9cfa-4d09-9672-40de98632e0e" (UID: "d3469c1a-9cfa-4d09-9672-40de98632e0e"). InnerVolumeSpecName "kube-api-access-h6txp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.895011 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3469c1a-9cfa-4d09-9672-40de98632e0e" (UID: "d3469c1a-9cfa-4d09-9672-40de98632e0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.919856 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-config-data" (OuterVolumeSpecName: "config-data") pod "d3469c1a-9cfa-4d09-9672-40de98632e0e" (UID: "d3469c1a-9cfa-4d09-9672-40de98632e0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.956382 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.956450 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3469c1a-9cfa-4d09-9672-40de98632e0e-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:13 crc kubenswrapper[4799]: I0930 14:39:13.956464 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6txp\" (UniqueName: \"kubernetes.io/projected/d3469c1a-9cfa-4d09-9672-40de98632e0e-kube-api-access-h6txp\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.406016 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nxr4c" event={"ID":"d3469c1a-9cfa-4d09-9672-40de98632e0e","Type":"ContainerDied","Data":"e111c33be1b0eb3f300da8866a85a988f53eec351bcdd8b68fa57f801a5d91fe"} Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.406079 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e111c33be1b0eb3f300da8866a85a988f53eec351bcdd8b68fa57f801a5d91fe" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.406136 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nxr4c" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.662293 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-s4n2b"] Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684014 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6zlj6"] Sep 30 14:39:14 crc kubenswrapper[4799]: E0930 14:39:14.684484 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f6895b-e518-4d47-8b9e-ad1143dbc433" containerName="init" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684504 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f6895b-e518-4d47-8b9e-ad1143dbc433" containerName="init" Sep 30 14:39:14 crc kubenswrapper[4799]: E0930 14:39:14.684520 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3469c1a-9cfa-4d09-9672-40de98632e0e" containerName="keystone-db-sync" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684527 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3469c1a-9cfa-4d09-9672-40de98632e0e" containerName="keystone-db-sync" Sep 30 14:39:14 crc kubenswrapper[4799]: E0930 14:39:14.684549 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea31ca1c-9064-41a9-a55e-24d09ee891f3" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684558 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea31ca1c-9064-41a9-a55e-24d09ee891f3" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: E0930 14:39:14.684585 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684593 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: E0930 14:39:14.684605 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f6895b-e518-4d47-8b9e-ad1143dbc433" containerName="dnsmasq-dns" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684612 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f6895b-e518-4d47-8b9e-ad1143dbc433" containerName="dnsmasq-dns" Sep 30 14:39:14 crc kubenswrapper[4799]: E0930 14:39:14.684620 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acec8647-568b-474c-86a6-590726b7b86e" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684626 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="acec8647-568b-474c-86a6-590726b7b86e" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684808 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="acec8647-568b-474c-86a6-590726b7b86e" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684824 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f6895b-e518-4d47-8b9e-ad1143dbc433" containerName="dnsmasq-dns" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684842 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3469c1a-9cfa-4d09-9672-40de98632e0e" containerName="keystone-db-sync" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684854 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea31ca1c-9064-41a9-a55e-24d09ee891f3" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.684861 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499" containerName="mariadb-account-create" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.685577 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.693067 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.693294 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.693422 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.693644 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vnkdb" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.742580 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6zlj6"] Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.774708 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-credential-keys\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.774801 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-scripts\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.774842 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-fernet-keys\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.774883 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqxn4\" (UniqueName: \"kubernetes.io/projected/ace1f710-ccd4-4825-8a6d-b66d508a217b-kube-api-access-rqxn4\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.774930 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-combined-ca-bundle\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.774956 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-config-data\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.780513 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr"] Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.788839 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.795901 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr"] Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877208 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-config\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877302 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-fernet-keys\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877344 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877399 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqxn4\" (UniqueName: \"kubernetes.io/projected/ace1f710-ccd4-4825-8a6d-b66d508a217b-kube-api-access-rqxn4\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877431 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-combined-ca-bundle\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877530 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-config-data\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877570 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-credential-keys\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877619 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877668 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9kf9\" (UniqueName: \"kubernetes.io/projected/27c6b8df-709e-4c81-87a5-6b59a2da1423-kube-api-access-k9kf9\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877863 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-scripts\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.877904 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.890763 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-fernet-keys\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.894849 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-config-data\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.901058 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-combined-ca-bundle\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.901432 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-credential-keys\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.915104 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-scripts\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.929438 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqxn4\" (UniqueName: \"kubernetes.io/projected/ace1f710-ccd4-4825-8a6d-b66d508a217b-kube-api-access-rqxn4\") pod \"keystone-bootstrap-6zlj6\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.985739 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.985822 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-config\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.985862 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.985912 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.986010 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.986033 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9kf9\" (UniqueName: \"kubernetes.io/projected/27c6b8df-709e-4c81-87a5-6b59a2da1423-kube-api-access-k9kf9\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.987459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.988102 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-config\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.988755 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.989890 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:14 crc kubenswrapper[4799]: I0930 14:39:14.996942 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.012491 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9kf9\" (UniqueName: \"kubernetes.io/projected/27c6b8df-709e-4c81-87a5-6b59a2da1423-kube-api-access-k9kf9\") pod \"dnsmasq-dns-5c5cc7c5ff-8zbnr\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.023161 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.140166 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.230039 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-fb472"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.234009 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.286341 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tzfkp" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.286705 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.292115 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.296704 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-combined-ca-bundle\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.296752 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-config\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.296873 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9459l\" (UniqueName: \"kubernetes.io/projected/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-kube-api-access-9459l\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.307299 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fb472"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.320032 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-576cc8ddf-zr27w"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.321691 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.327871 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.327945 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.328094 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-fp4t9" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.328446 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.398964 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rch4\" (UniqueName: \"kubernetes.io/projected/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-kube-api-access-5rch4\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.399353 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-horizon-secret-key\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.399416 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9459l\" (UniqueName: \"kubernetes.io/projected/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-kube-api-access-9459l\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.399455 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-scripts\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.399495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-combined-ca-bundle\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.399525 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-config\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.399549 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-config-data\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.399579 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-logs\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.407330 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-combined-ca-bundle\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.412371 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-576cc8ddf-zr27w"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.429571 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-config\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.430488 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" podUID="bdb12c35-9f78-4995-b550-573981289693" containerName="dnsmasq-dns" containerID="cri-o://1fdbd0951f319d68c563518077053a1e2f091c7f8100221c9286ff24c460b152" gracePeriod=10 Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.452114 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qrs4k"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.453494 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.466372 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-j4x5t" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.466673 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.466836 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.467409 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9459l\" (UniqueName: \"kubernetes.io/projected/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-kube-api-access-9459l\") pod \"neutron-db-sync-fb472\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.494946 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.500877 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-combined-ca-bundle\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.500936 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rch4\" (UniqueName: \"kubernetes.io/projected/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-kube-api-access-5rch4\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.500987 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-horizon-secret-key\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.501044 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-scripts\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.501079 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-config-data\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.501148 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-config-data\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.501181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-logs\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.501216 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71e7667-1d26-47fc-9884-1c946e72c244-logs\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.501245 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpqzs\" (UniqueName: \"kubernetes.io/projected/e71e7667-1d26-47fc-9884-1c946e72c244-kube-api-access-rpqzs\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.501298 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-scripts\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.502431 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-scripts\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.502902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-logs\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.509945 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-config-data\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.518285 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-horizon-secret-key\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.575848 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rch4\" (UniqueName: \"kubernetes.io/projected/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-kube-api-access-5rch4\") pod \"horizon-576cc8ddf-zr27w\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.603841 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-combined-ca-bundle\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.603922 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-config-data\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.603962 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71e7667-1d26-47fc-9884-1c946e72c244-logs\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.603981 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpqzs\" (UniqueName: \"kubernetes.io/projected/e71e7667-1d26-47fc-9884-1c946e72c244-kube-api-access-rpqzs\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.604015 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-scripts\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.606323 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71e7667-1d26-47fc-9884-1c946e72c244-logs\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.606741 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fb472" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.615601 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qrs4k"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.657488 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-scripts\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.659049 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.683471 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-config-data\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.684345 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-combined-ca-bundle\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.699839 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpqzs\" (UniqueName: \"kubernetes.io/projected/e71e7667-1d26-47fc-9884-1c946e72c244-kube-api-access-rpqzs\") pod \"placement-db-sync-qrs4k\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.820940 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-mcsxn"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.836293 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qrs4k" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.846856 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.957832 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-mcsxn"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.968197 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85c8f46c9-sxlj2"] Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.970440 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:15 crc kubenswrapper[4799]: I0930 14:39:15.975221 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85c8f46c9-sxlj2"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.031323 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.034557 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.046917 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxff\" (UniqueName: \"kubernetes.io/projected/89554c30-2812-4e26-9878-a972ab70dd7e-kube-api-access-cpxff\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.051896 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.052252 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.052425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.052526 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.052693 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-config\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.073545 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.075916 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.086066 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.148560 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.164980 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.165020 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pjthr" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.166234 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195158 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.167125 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.166133 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195321 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx7zv\" (UniqueName: \"kubernetes.io/projected/cd86e414-219a-4754-85c4-239ef97843c1-kube-api-access-wx7zv\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195393 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195431 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195472 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-config\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.166183 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195537 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8nrj\" (UniqueName: \"kubernetes.io/projected/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-kube-api-access-f8nrj\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195586 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-scripts\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195641 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-run-httpd\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195719 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-scripts\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195752 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-config-data\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195815 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-config-data\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195842 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-log-httpd\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195941 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd86e414-219a-4754-85c4-239ef97843c1-horizon-secret-key\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.195996 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.196057 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxff\" (UniqueName: \"kubernetes.io/projected/89554c30-2812-4e26-9878-a972ab70dd7e-kube-api-access-cpxff\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.196091 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd86e414-219a-4754-85c4-239ef97843c1-logs\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.196121 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.197670 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-config\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.198270 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.198818 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.204510 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.255803 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298284 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-config-data\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298357 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-log-httpd\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298407 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-logs\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298450 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298486 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd86e414-219a-4754-85c4-239ef97843c1-horizon-secret-key\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298538 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298561 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298623 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd86e414-219a-4754-85c4-239ef97843c1-logs\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298666 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-config-data\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298704 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298741 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx7zv\" (UniqueName: \"kubernetes.io/projected/cd86e414-219a-4754-85c4-239ef97843c1-kube-api-access-wx7zv\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298799 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8nrj\" (UniqueName: \"kubernetes.io/projected/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-kube-api-access-f8nrj\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298821 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298853 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-scripts\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298878 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-scripts\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298913 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-run-httpd\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298944 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r7kc\" (UniqueName: \"kubernetes.io/projected/851d4e8d-09ef-4acf-b813-c3b715ff54b2-kube-api-access-6r7kc\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298972 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.298993 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-scripts\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.299016 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-config-data\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.307188 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-config-data\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.308414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-run-httpd\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.308794 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-scripts\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.309266 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd86e414-219a-4754-85c4-239ef97843c1-logs\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.309845 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-log-httpd\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.317868 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxff\" (UniqueName: \"kubernetes.io/projected/89554c30-2812-4e26-9878-a972ab70dd7e-kube-api-access-cpxff\") pod \"dnsmasq-dns-8b5c85b87-mcsxn\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.318530 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd86e414-219a-4754-85c4-239ef97843c1-horizon-secret-key\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.319259 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.325828 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.341396 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-scripts\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.342186 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-config-data\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.343011 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.355229 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.361842 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.364127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx7zv\" (UniqueName: \"kubernetes.io/projected/cd86e414-219a-4754-85c4-239ef97843c1-kube-api-access-wx7zv\") pod \"horizon-85c8f46c9-sxlj2\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.367764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8nrj\" (UniqueName: \"kubernetes.io/projected/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-kube-api-access-f8nrj\") pod \"ceilometer-0\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.378399 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.378523 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.386234 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400367 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r7kc\" (UniqueName: \"kubernetes.io/projected/851d4e8d-09ef-4acf-b813-c3b715ff54b2-kube-api-access-6r7kc\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400416 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400449 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400473 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5ds8\" (UniqueName: \"kubernetes.io/projected/b029d8e7-528a-47cb-9e1d-20c165035dfc-kube-api-access-h5ds8\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400516 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-logs\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400539 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-logs\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400556 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400582 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400599 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400671 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400709 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400726 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-config-data\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400751 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400769 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400816 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.400856 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-scripts\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.404368 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.404424 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-logs\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.416914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.423444 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.430943 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.447641 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-scripts\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.454243 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.459050 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.461554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-config-data\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.461675 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.505083 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.505132 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5ds8\" (UniqueName: \"kubernetes.io/projected/b029d8e7-528a-47cb-9e1d-20c165035dfc-kube-api-access-h5ds8\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.505171 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-logs\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.505199 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.505231 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.505281 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.505308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.505339 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.511813 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.517088 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-logs\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.532038 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.556683 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r7kc\" (UniqueName: \"kubernetes.io/projected/851d4e8d-09ef-4acf-b813-c3b715ff54b2-kube-api-access-6r7kc\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.578388 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.601879 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.630901 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.641116 4799 generic.go:334] "Generic (PLEG): container finished" podID="bdb12c35-9f78-4995-b550-573981289693" containerID="1fdbd0951f319d68c563518077053a1e2f091c7f8100221c9286ff24c460b152" exitCode=0 Sep 30 14:39:16 crc kubenswrapper[4799]: E0930 14:39:16.660858 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data glance internal-tls-certs kube-api-access-h5ds8], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="b029d8e7-528a-47cb-9e1d-20c165035dfc" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.680575 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5ds8\" (UniqueName: \"kubernetes.io/projected/b029d8e7-528a-47cb-9e1d-20c165035dfc-kube-api-access-h5ds8\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.682621 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.685470 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.687429 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.708173 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" event={"ID":"bdb12c35-9f78-4995-b550-573981289693","Type":"ContainerDied","Data":"1fdbd0951f319d68c563518077053a1e2f091c7f8100221c9286ff24c460b152"} Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.716601 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.770425 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6zlj6"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.852432 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:16 crc kubenswrapper[4799]: I0930 14:39:16.853386 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.262391 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr"] Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.673176 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6zlj6" event={"ID":"ace1f710-ccd4-4825-8a6d-b66d508a217b","Type":"ContainerStarted","Data":"f3c68016490a25dd53bd16f1573912f6921288ae36e5a7e6fb85b6fb0af37ad8"} Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.675311 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.676273 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" event={"ID":"27c6b8df-709e-4c81-87a5-6b59a2da1423","Type":"ContainerStarted","Data":"bd6b49bd0f8a9b8df8fdd005086031d78ddfc827c5faa6d2f2edae3c06772425"} Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.717865 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.741101 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-mx9fz"] Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.742806 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.748512 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gz9nw" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.749208 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.750753 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-combined-ca-bundle\") pod \"b029d8e7-528a-47cb-9e1d-20c165035dfc\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.750884 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-httpd-run\") pod \"b029d8e7-528a-47cb-9e1d-20c165035dfc\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.750993 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-logs\") pod \"b029d8e7-528a-47cb-9e1d-20c165035dfc\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.751092 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-config-data\") pod \"b029d8e7-528a-47cb-9e1d-20c165035dfc\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.751204 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-scripts\") pod \"b029d8e7-528a-47cb-9e1d-20c165035dfc\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.751316 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-internal-tls-certs\") pod \"b029d8e7-528a-47cb-9e1d-20c165035dfc\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.751396 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"b029d8e7-528a-47cb-9e1d-20c165035dfc\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.751516 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5ds8\" (UniqueName: \"kubernetes.io/projected/b029d8e7-528a-47cb-9e1d-20c165035dfc-kube-api-access-h5ds8\") pod \"b029d8e7-528a-47cb-9e1d-20c165035dfc\" (UID: \"b029d8e7-528a-47cb-9e1d-20c165035dfc\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.770092 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2txv\" (UniqueName: \"kubernetes.io/projected/78ac655a-3fc7-46a4-93c2-69574e8b8f22-kube-api-access-k2txv\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.770215 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-scripts\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.771120 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-config-data\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.771177 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-combined-ca-bundle\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.771232 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-db-sync-config-data\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.771276 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ac655a-3fc7-46a4-93c2-69574e8b8f22-etc-machine-id\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.762838 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.771908 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b029d8e7-528a-47cb-9e1d-20c165035dfc" (UID: "b029d8e7-528a-47cb-9e1d-20c165035dfc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.772152 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-logs" (OuterVolumeSpecName: "logs") pod "b029d8e7-528a-47cb-9e1d-20c165035dfc" (UID: "b029d8e7-528a-47cb-9e1d-20c165035dfc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.783896 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-scripts" (OuterVolumeSpecName: "scripts") pod "b029d8e7-528a-47cb-9e1d-20c165035dfc" (UID: "b029d8e7-528a-47cb-9e1d-20c165035dfc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.788755 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-mx9fz"] Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.855684 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b029d8e7-528a-47cb-9e1d-20c165035dfc-kube-api-access-h5ds8" (OuterVolumeSpecName: "kube-api-access-h5ds8") pod "b029d8e7-528a-47cb-9e1d-20c165035dfc" (UID: "b029d8e7-528a-47cb-9e1d-20c165035dfc"). InnerVolumeSpecName "kube-api-access-h5ds8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.857010 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b029d8e7-528a-47cb-9e1d-20c165035dfc" (UID: "b029d8e7-528a-47cb-9e1d-20c165035dfc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-scripts\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-config-data\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-combined-ca-bundle\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873619 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-db-sync-config-data\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873642 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ac655a-3fc7-46a4-93c2-69574e8b8f22-etc-machine-id\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873769 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2txv\" (UniqueName: \"kubernetes.io/projected/78ac655a-3fc7-46a4-93c2-69574e8b8f22-kube-api-access-k2txv\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873868 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873882 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873895 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5ds8\" (UniqueName: \"kubernetes.io/projected/b029d8e7-528a-47cb-9e1d-20c165035dfc-kube-api-access-h5ds8\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873903 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.873913 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b029d8e7-528a-47cb-9e1d-20c165035dfc-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.882427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ac655a-3fc7-46a4-93c2-69574e8b8f22-etc-machine-id\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.890562 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "b029d8e7-528a-47cb-9e1d-20c165035dfc" (UID: "b029d8e7-528a-47cb-9e1d-20c165035dfc"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.899159 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-combined-ca-bundle\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.903127 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b029d8e7-528a-47cb-9e1d-20c165035dfc" (UID: "b029d8e7-528a-47cb-9e1d-20c165035dfc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.904728 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-config-data" (OuterVolumeSpecName: "config-data") pod "b029d8e7-528a-47cb-9e1d-20c165035dfc" (UID: "b029d8e7-528a-47cb-9e1d-20c165035dfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.904780 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.905555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-config-data\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.906456 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-db-sync-config-data\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.933400 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2txv\" (UniqueName: \"kubernetes.io/projected/78ac655a-3fc7-46a4-93c2-69574e8b8f22-kube-api-access-k2txv\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.947632 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-scripts\") pod \"cinder-db-sync-mx9fz\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.978110 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqzdz\" (UniqueName: \"kubernetes.io/projected/bdb12c35-9f78-4995-b550-573981289693-kube-api-access-zqzdz\") pod \"bdb12c35-9f78-4995-b550-573981289693\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.978267 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-swift-storage-0\") pod \"bdb12c35-9f78-4995-b550-573981289693\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.978339 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-svc\") pod \"bdb12c35-9f78-4995-b550-573981289693\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.978399 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-config\") pod \"bdb12c35-9f78-4995-b550-573981289693\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.978476 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-sb\") pod \"bdb12c35-9f78-4995-b550-573981289693\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.978533 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-nb\") pod \"bdb12c35-9f78-4995-b550-573981289693\" (UID: \"bdb12c35-9f78-4995-b550-573981289693\") " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.979057 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.979090 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Sep 30 14:39:17 crc kubenswrapper[4799]: I0930 14:39:17.979104 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b029d8e7-528a-47cb-9e1d-20c165035dfc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.029214 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xt6j5"] Sep 30 14:39:18 crc kubenswrapper[4799]: E0930 14:39:18.029814 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb12c35-9f78-4995-b550-573981289693" containerName="init" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.029840 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb12c35-9f78-4995-b550-573981289693" containerName="init" Sep 30 14:39:18 crc kubenswrapper[4799]: E0930 14:39:18.029854 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb12c35-9f78-4995-b550-573981289693" containerName="dnsmasq-dns" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.029859 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb12c35-9f78-4995-b550-573981289693" containerName="dnsmasq-dns" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.030058 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdb12c35-9f78-4995-b550-573981289693" containerName="dnsmasq-dns" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.041983 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.059737 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8vn4g" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.060260 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.070834 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdb12c35-9f78-4995-b550-573981289693-kube-api-access-zqzdz" (OuterVolumeSpecName: "kube-api-access-zqzdz") pod "bdb12c35-9f78-4995-b550-573981289693" (UID: "bdb12c35-9f78-4995-b550-573981289693"). InnerVolumeSpecName "kube-api-access-zqzdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.080883 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tznlh\" (UniqueName: \"kubernetes.io/projected/2d30b6ec-a2d1-4f93-932b-e78856253cc9-kube-api-access-tznlh\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.081301 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-db-sync-config-data\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.081408 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-combined-ca-bundle\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.081608 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqzdz\" (UniqueName: \"kubernetes.io/projected/bdb12c35-9f78-4995-b550-573981289693-kube-api-access-zqzdz\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.089112 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xt6j5"] Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.117868 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.173462 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.189056 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-combined-ca-bundle\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.189199 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tznlh\" (UniqueName: \"kubernetes.io/projected/2d30b6ec-a2d1-4f93-932b-e78856253cc9-kube-api-access-tznlh\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.189222 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-db-sync-config-data\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.189293 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.206242 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-db-sync-config-data\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.206728 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fb472"] Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.212901 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-combined-ca-bundle\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.232857 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-576cc8ddf-zr27w"] Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.245069 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qrs4k"] Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.250090 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tznlh\" (UniqueName: \"kubernetes.io/projected/2d30b6ec-a2d1-4f93-932b-e78856253cc9-kube-api-access-tznlh\") pod \"barbican-db-sync-xt6j5\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.462905 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-mcsxn"] Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.532298 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bdb12c35-9f78-4995-b550-573981289693" (UID: "bdb12c35-9f78-4995-b550-573981289693"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.613064 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85c8f46c9-sxlj2"] Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.651884 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.669696 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-config" (OuterVolumeSpecName: "config") pod "bdb12c35-9f78-4995-b550-573981289693" (UID: "bdb12c35-9f78-4995-b550-573981289693"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.671116 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.672798 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bdb12c35-9f78-4995-b550-573981289693" (UID: "bdb12c35-9f78-4995-b550-573981289693"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.706587 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bdb12c35-9f78-4995-b550-573981289693" (UID: "bdb12c35-9f78-4995-b550-573981289693"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.711729 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bdb12c35-9f78-4995-b550-573981289693" (UID: "bdb12c35-9f78-4995-b550-573981289693"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.751272 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85c8f46c9-sxlj2" event={"ID":"cd86e414-219a-4754-85c4-239ef97843c1","Type":"ContainerStarted","Data":"8f4fcbbfb06ee03188e82cda7ec55ee9c4cab3bf2ea599fe8c926ad469c5ee2f"} Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.766495 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-576cc8ddf-zr27w" event={"ID":"c78ce9ff-e63b-4587-8943-daf6fe7b4b82","Type":"ContainerStarted","Data":"9a5287604946212723bc5832648dcd12c158eb9db39e0235b4c9ebf29aec24e8"} Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.771795 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" event={"ID":"89554c30-2812-4e26-9878-a972ab70dd7e","Type":"ContainerStarted","Data":"c6e433ba9151efb327a6b5fcfdb18a5bdef822227e1b39e36e947be55b926178"} Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.785379 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.786112 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.786904 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.788246 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bdb12c35-9f78-4995-b550-573981289693-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.797406 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6zlj6" event={"ID":"ace1f710-ccd4-4825-8a6d-b66d508a217b","Type":"ContainerStarted","Data":"f5239377f75f41afe5178e8987a478a2f4461fff97d01826fa1359713ddf2fce"} Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.816674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fb472" event={"ID":"67e3be2e-5fd3-4ac7-9138-c968d3d0c025","Type":"ContainerStarted","Data":"86b357cf5489ea254e95d900f8ca9196b26ab86b59ac53354f0840bb11a666f1"} Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.877122 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qrs4k" event={"ID":"e71e7667-1d26-47fc-9884-1c946e72c244","Type":"ContainerStarted","Data":"4b7b993575c7d3ebbc1d07e25abb8a09e48284374ae9ad79f7c1c184672ede8a"} Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.906173 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.906448 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-s4n2b" event={"ID":"bdb12c35-9f78-4995-b550-573981289693","Type":"ContainerDied","Data":"a24a8ad4d1dd96357d41fedc66a5211355fb104bf59b74a10fc8eeff7e1c443d"} Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.906509 4799 scope.go:117] "RemoveContainer" containerID="1fdbd0951f319d68c563518077053a1e2f091c7f8100221c9286ff24c460b152" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.916432 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.917237 4799 generic.go:334] "Generic (PLEG): container finished" podID="27c6b8df-709e-4c81-87a5-6b59a2da1423" containerID="a7f6a39262c9f8daf39495afe60161532f01e7728e4313fb6b403b7165dc0f6f" exitCode=0 Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.917378 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.919522 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" event={"ID":"27c6b8df-709e-4c81-87a5-6b59a2da1423","Type":"ContainerDied","Data":"a7f6a39262c9f8daf39495afe60161532f01e7728e4313fb6b403b7165dc0f6f"} Sep 30 14:39:18 crc kubenswrapper[4799]: I0930 14:39:18.944421 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6zlj6" podStartSLOduration=4.944393258 podStartE2EDuration="4.944393258s" podCreationTimestamp="2025-09-30 14:39:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:18.85419263 +0000 UTC m=+1180.937793067" watchObservedRunningTime="2025-09-30 14:39:18.944393258 +0000 UTC m=+1181.027993685" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.095583 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.134700 4799 scope.go:117] "RemoveContainer" containerID="0ce0147026b273b9297927ca0054d3adcd9b22e55cdd543329a1e06d5193e758" Sep 30 14:39:19 crc kubenswrapper[4799]: W0930 14:39:19.295641 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78ac655a_3fc7_46a4_93c2_69574e8b8f22.slice/crio-3f86766104c25656916fcf4c1cfadbc141c5990594f7857b7ef2f2c800149e12 WatchSource:0}: Error finding container 3f86766104c25656916fcf4c1cfadbc141c5990594f7857b7ef2f2c800149e12: Status 404 returned error can't find the container with id 3f86766104c25656916fcf4c1cfadbc141c5990594f7857b7ef2f2c800149e12 Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.316145 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-mx9fz"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.393187 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.419964 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.451489 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-s4n2b"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.458366 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-s4n2b"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.466738 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.468813 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.475135 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.475213 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.476540 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.610839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.610958 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.610996 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.611022 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.611045 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.611067 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-logs\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.611083 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.611114 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csj2f\" (UniqueName: \"kubernetes.io/projected/67d7038d-99dc-47b8-9a91-67df94bcb35f-kube-api-access-csj2f\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.694300 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xt6j5"] Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.713283 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.713398 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.713433 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.713456 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.713479 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.713504 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-logs\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.713539 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.713570 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csj2f\" (UniqueName: \"kubernetes.io/projected/67d7038d-99dc-47b8-9a91-67df94bcb35f-kube-api-access-csj2f\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.715563 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.716246 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.719452 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-logs\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.743026 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.751347 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.759340 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.764629 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.765523 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csj2f\" (UniqueName: \"kubernetes.io/projected/67d7038d-99dc-47b8-9a91-67df94bcb35f-kube-api-access-csj2f\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.788070 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:19 crc kubenswrapper[4799]: I0930 14:39:19.847011 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.071430 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"851d4e8d-09ef-4acf-b813-c3b715ff54b2","Type":"ContainerStarted","Data":"84b7124feb22b045291a22c75d1bc62fa2a11f83ae62531aa6f412b7a3831e12"} Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.129980 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xt6j5" event={"ID":"2d30b6ec-a2d1-4f93-932b-e78856253cc9","Type":"ContainerStarted","Data":"9b2ed069427d7da7cd9be63211e1d604a39fdf28d1b95686cb02474f506e43f4"} Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.159452 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc98dba3-94f3-4cc0-91b4-152a3fa006b7","Type":"ContainerStarted","Data":"732391f9f2193e29ff38a4a58268c34c98aad094fc22cc7167811aaf3fdfc5f4"} Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.236594 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fb472" event={"ID":"67e3be2e-5fd3-4ac7-9138-c968d3d0c025","Type":"ContainerStarted","Data":"85994786d1236ad433f7cb78913926fed48b7e24160914eb8d1d855e9768c479"} Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.253099 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-mx9fz" event={"ID":"78ac655a-3fc7-46a4-93c2-69574e8b8f22","Type":"ContainerStarted","Data":"3f86766104c25656916fcf4c1cfadbc141c5990594f7857b7ef2f2c800149e12"} Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.333270 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-fb472" podStartSLOduration=5.306418515 podStartE2EDuration="5.306418515s" podCreationTimestamp="2025-09-30 14:39:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:20.284835774 +0000 UTC m=+1182.368436211" watchObservedRunningTime="2025-09-30 14:39:20.306418515 +0000 UTC m=+1182.390018952" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.468828 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.558871 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b029d8e7-528a-47cb-9e1d-20c165035dfc" path="/var/lib/kubelet/pods/b029d8e7-528a-47cb-9e1d-20c165035dfc/volumes" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.572245 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdb12c35-9f78-4995-b550-573981289693" path="/var/lib/kubelet/pods/bdb12c35-9f78-4995-b550-573981289693/volumes" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.640907 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-config\") pod \"27c6b8df-709e-4c81-87a5-6b59a2da1423\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.641352 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-swift-storage-0\") pod \"27c6b8df-709e-4c81-87a5-6b59a2da1423\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.641396 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-svc\") pod \"27c6b8df-709e-4c81-87a5-6b59a2da1423\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.641452 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9kf9\" (UniqueName: \"kubernetes.io/projected/27c6b8df-709e-4c81-87a5-6b59a2da1423-kube-api-access-k9kf9\") pod \"27c6b8df-709e-4c81-87a5-6b59a2da1423\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.641571 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-nb\") pod \"27c6b8df-709e-4c81-87a5-6b59a2da1423\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.641777 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-sb\") pod \"27c6b8df-709e-4c81-87a5-6b59a2da1423\" (UID: \"27c6b8df-709e-4c81-87a5-6b59a2da1423\") " Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.733523 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27c6b8df-709e-4c81-87a5-6b59a2da1423-kube-api-access-k9kf9" (OuterVolumeSpecName: "kube-api-access-k9kf9") pod "27c6b8df-709e-4c81-87a5-6b59a2da1423" (UID: "27c6b8df-709e-4c81-87a5-6b59a2da1423"). InnerVolumeSpecName "kube-api-access-k9kf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.778070 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "27c6b8df-709e-4c81-87a5-6b59a2da1423" (UID: "27c6b8df-709e-4c81-87a5-6b59a2da1423"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.779269 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-config" (OuterVolumeSpecName: "config") pod "27c6b8df-709e-4c81-87a5-6b59a2da1423" (UID: "27c6b8df-709e-4c81-87a5-6b59a2da1423"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.865199 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.865266 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9kf9\" (UniqueName: \"kubernetes.io/projected/27c6b8df-709e-4c81-87a5-6b59a2da1423-kube-api-access-k9kf9\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.865278 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.988806 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "27c6b8df-709e-4c81-87a5-6b59a2da1423" (UID: "27c6b8df-709e-4c81-87a5-6b59a2da1423"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:20 crc kubenswrapper[4799]: I0930 14:39:20.991238 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "27c6b8df-709e-4c81-87a5-6b59a2da1423" (UID: "27c6b8df-709e-4c81-87a5-6b59a2da1423"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.012221 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "27c6b8df-709e-4c81-87a5-6b59a2da1423" (UID: "27c6b8df-709e-4c81-87a5-6b59a2da1423"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.059905 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-576cc8ddf-zr27w"] Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.073811 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.088774 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77997c977-fsfdt"] Sep 30 14:39:21 crc kubenswrapper[4799]: E0930 14:39:21.089220 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c6b8df-709e-4c81-87a5-6b59a2da1423" containerName="init" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.089234 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c6b8df-709e-4c81-87a5-6b59a2da1423" containerName="init" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.089443 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="27c6b8df-709e-4c81-87a5-6b59a2da1423" containerName="init" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.091407 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.091447 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.091458 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27c6b8df-709e-4c81-87a5-6b59a2da1423-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.091686 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.111953 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77997c977-fsfdt"] Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.155353 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.193135 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-horizon-secret-key\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.193239 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-scripts\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.193268 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcknk\" (UniqueName: \"kubernetes.io/projected/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-kube-api-access-jcknk\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.193333 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-logs\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.193373 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-config-data\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.295925 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-scripts\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.296260 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcknk\" (UniqueName: \"kubernetes.io/projected/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-kube-api-access-jcknk\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.296343 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-logs\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.296394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-config-data\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.296446 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-horizon-secret-key\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.298366 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-scripts\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.299639 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-logs\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.299790 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-config-data\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.309196 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" event={"ID":"27c6b8df-709e-4c81-87a5-6b59a2da1423","Type":"ContainerDied","Data":"bd6b49bd0f8a9b8df8fdd005086031d78ddfc827c5faa6d2f2edae3c06772425"} Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.309271 4799 scope.go:117] "RemoveContainer" containerID="a7f6a39262c9f8daf39495afe60161532f01e7728e4313fb6b403b7165dc0f6f" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.309448 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.320063 4799 generic.go:334] "Generic (PLEG): container finished" podID="89554c30-2812-4e26-9878-a972ab70dd7e" containerID="99350693c10131040351e9357be558d8cd11740ed3bebb18ce0bf8d2bda72471" exitCode=0 Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.321874 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" event={"ID":"89554c30-2812-4e26-9878-a972ab70dd7e","Type":"ContainerDied","Data":"99350693c10131040351e9357be558d8cd11740ed3bebb18ce0bf8d2bda72471"} Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.343633 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-horizon-secret-key\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.344130 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcknk\" (UniqueName: \"kubernetes.io/projected/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-kube-api-access-jcknk\") pod \"horizon-77997c977-fsfdt\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.440891 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.465738 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.555931 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr"] Sep 30 14:39:21 crc kubenswrapper[4799]: W0930 14:39:21.563575 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d7038d_99dc_47b8_9a91_67df94bcb35f.slice/crio-128746b890dbcd009cc49b08b31e01a51c0871cdb8fa38307e310f6ca34f0b20 WatchSource:0}: Error finding container 128746b890dbcd009cc49b08b31e01a51c0871cdb8fa38307e310f6ca34f0b20: Status 404 returned error can't find the container with id 128746b890dbcd009cc49b08b31e01a51c0871cdb8fa38307e310f6ca34f0b20 Sep 30 14:39:21 crc kubenswrapper[4799]: I0930 14:39:21.565434 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-8zbnr"] Sep 30 14:39:22 crc kubenswrapper[4799]: I0930 14:39:22.404176 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"851d4e8d-09ef-4acf-b813-c3b715ff54b2","Type":"ContainerStarted","Data":"57c7031a2996e8bdded23e20b9ae624ead1c3c9db9e31352b6b09c69a5b5f18d"} Sep 30 14:39:22 crc kubenswrapper[4799]: I0930 14:39:22.414731 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67d7038d-99dc-47b8-9a91-67df94bcb35f","Type":"ContainerStarted","Data":"128746b890dbcd009cc49b08b31e01a51c0871cdb8fa38307e310f6ca34f0b20"} Sep 30 14:39:22 crc kubenswrapper[4799]: I0930 14:39:22.599187 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27c6b8df-709e-4c81-87a5-6b59a2da1423" path="/var/lib/kubelet/pods/27c6b8df-709e-4c81-87a5-6b59a2da1423/volumes" Sep 30 14:39:22 crc kubenswrapper[4799]: I0930 14:39:22.671251 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77997c977-fsfdt"] Sep 30 14:39:23 crc kubenswrapper[4799]: I0930 14:39:23.444689 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" event={"ID":"89554c30-2812-4e26-9878-a972ab70dd7e","Type":"ContainerStarted","Data":"fcfbbd41979a3563f6feebb2661150b16ce97a70f3f9e5bb53e5e85e4976b6d1"} Sep 30 14:39:23 crc kubenswrapper[4799]: I0930 14:39:23.447355 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77997c977-fsfdt" event={"ID":"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8","Type":"ContainerStarted","Data":"1257a900b53f4751acbecd072d09c4012a5db5c27ee334003d8e84c5501df266"} Sep 30 14:39:23 crc kubenswrapper[4799]: I0930 14:39:23.447616 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:24 crc kubenswrapper[4799]: I0930 14:39:24.542595 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerName="glance-log" containerID="cri-o://57c7031a2996e8bdded23e20b9ae624ead1c3c9db9e31352b6b09c69a5b5f18d" gracePeriod=30 Sep 30 14:39:24 crc kubenswrapper[4799]: I0930 14:39:24.543902 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerName="glance-httpd" containerID="cri-o://3472d04a16319ea3e79e3e0307f153f6f987daa35cdb088bfc332bdebaa60225" gracePeriod=30 Sep 30 14:39:24 crc kubenswrapper[4799]: I0930 14:39:24.586407 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" podStartSLOduration=9.580756661 podStartE2EDuration="9.580756661s" podCreationTimestamp="2025-09-30 14:39:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:23.480285899 +0000 UTC m=+1185.563886346" watchObservedRunningTime="2025-09-30 14:39:24.580756661 +0000 UTC m=+1186.664357088" Sep 30 14:39:24 crc kubenswrapper[4799]: I0930 14:39:24.587785 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.587761305 podStartE2EDuration="10.587761305s" podCreationTimestamp="2025-09-30 14:39:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:24.573954282 +0000 UTC m=+1186.657554709" watchObservedRunningTime="2025-09-30 14:39:24.587761305 +0000 UTC m=+1186.671361732" Sep 30 14:39:24 crc kubenswrapper[4799]: I0930 14:39:24.587816 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67d7038d-99dc-47b8-9a91-67df94bcb35f","Type":"ContainerStarted","Data":"623dcd6a3e46baf1a0ff371edd5775b5e1092f41a9a55d17fa8b660b1166c953"} Sep 30 14:39:24 crc kubenswrapper[4799]: I0930 14:39:24.587851 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"851d4e8d-09ef-4acf-b813-c3b715ff54b2","Type":"ContainerStarted","Data":"3472d04a16319ea3e79e3e0307f153f6f987daa35cdb088bfc332bdebaa60225"} Sep 30 14:39:25 crc kubenswrapper[4799]: I0930 14:39:25.574701 4799 generic.go:334] "Generic (PLEG): container finished" podID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerID="3472d04a16319ea3e79e3e0307f153f6f987daa35cdb088bfc332bdebaa60225" exitCode=0 Sep 30 14:39:25 crc kubenswrapper[4799]: I0930 14:39:25.575057 4799 generic.go:334] "Generic (PLEG): container finished" podID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerID="57c7031a2996e8bdded23e20b9ae624ead1c3c9db9e31352b6b09c69a5b5f18d" exitCode=143 Sep 30 14:39:25 crc kubenswrapper[4799]: I0930 14:39:25.575091 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"851d4e8d-09ef-4acf-b813-c3b715ff54b2","Type":"ContainerDied","Data":"3472d04a16319ea3e79e3e0307f153f6f987daa35cdb088bfc332bdebaa60225"} Sep 30 14:39:25 crc kubenswrapper[4799]: I0930 14:39:25.575131 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"851d4e8d-09ef-4acf-b813-c3b715ff54b2","Type":"ContainerDied","Data":"57c7031a2996e8bdded23e20b9ae624ead1c3c9db9e31352b6b09c69a5b5f18d"} Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.373581 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.487529 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-combined-ca-bundle\") pod \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.487690 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-httpd-run\") pod \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.487829 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-public-tls-certs\") pod \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.487873 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-logs\") pod \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.487982 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.488029 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-config-data\") pod \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.488084 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r7kc\" (UniqueName: \"kubernetes.io/projected/851d4e8d-09ef-4acf-b813-c3b715ff54b2-kube-api-access-6r7kc\") pod \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.488112 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-scripts\") pod \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\" (UID: \"851d4e8d-09ef-4acf-b813-c3b715ff54b2\") " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.493305 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-logs" (OuterVolumeSpecName: "logs") pod "851d4e8d-09ef-4acf-b813-c3b715ff54b2" (UID: "851d4e8d-09ef-4acf-b813-c3b715ff54b2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.493795 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "851d4e8d-09ef-4acf-b813-c3b715ff54b2" (UID: "851d4e8d-09ef-4acf-b813-c3b715ff54b2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.512095 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "851d4e8d-09ef-4acf-b813-c3b715ff54b2" (UID: "851d4e8d-09ef-4acf-b813-c3b715ff54b2"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.512315 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/851d4e8d-09ef-4acf-b813-c3b715ff54b2-kube-api-access-6r7kc" (OuterVolumeSpecName: "kube-api-access-6r7kc") pod "851d4e8d-09ef-4acf-b813-c3b715ff54b2" (UID: "851d4e8d-09ef-4acf-b813-c3b715ff54b2"). InnerVolumeSpecName "kube-api-access-6r7kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.556329 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-scripts" (OuterVolumeSpecName: "scripts") pod "851d4e8d-09ef-4acf-b813-c3b715ff54b2" (UID: "851d4e8d-09ef-4acf-b813-c3b715ff54b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.608548 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r7kc\" (UniqueName: \"kubernetes.io/projected/851d4e8d-09ef-4acf-b813-c3b715ff54b2-kube-api-access-6r7kc\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.608583 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.608593 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.608603 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/851d4e8d-09ef-4acf-b813-c3b715ff54b2-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.608626 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.678857 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.698514 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67d7038d-99dc-47b8-9a91-67df94bcb35f","Type":"ContainerStarted","Data":"2561537a6d88c97f0e979a2ca7be6ea1e4f84566d5aabd02607c41b15c723eb3"} Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.698817 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerName="glance-log" containerID="cri-o://623dcd6a3e46baf1a0ff371edd5775b5e1092f41a9a55d17fa8b660b1166c953" gracePeriod=30 Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.699273 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerName="glance-httpd" containerID="cri-o://2561537a6d88c97f0e979a2ca7be6ea1e4f84566d5aabd02607c41b15c723eb3" gracePeriod=30 Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.716227 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.756185 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "851d4e8d-09ef-4acf-b813-c3b715ff54b2" (UID: "851d4e8d-09ef-4acf-b813-c3b715ff54b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.759118 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "851d4e8d-09ef-4acf-b813-c3b715ff54b2" (UID: "851d4e8d-09ef-4acf-b813-c3b715ff54b2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.764941 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.764912682 podStartE2EDuration="7.764912682s" podCreationTimestamp="2025-09-30 14:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:39:26.73679815 +0000 UTC m=+1188.820398587" watchObservedRunningTime="2025-09-30 14:39:26.764912682 +0000 UTC m=+1188.848513119" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.781965 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"851d4e8d-09ef-4acf-b813-c3b715ff54b2","Type":"ContainerDied","Data":"84b7124feb22b045291a22c75d1bc62fa2a11f83ae62531aa6f412b7a3831e12"} Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.782119 4799 scope.go:117] "RemoveContainer" containerID="3472d04a16319ea3e79e3e0307f153f6f987daa35cdb088bfc332bdebaa60225" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.783327 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.784186 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-config-data" (OuterVolumeSpecName: "config-data") pod "851d4e8d-09ef-4acf-b813-c3b715ff54b2" (UID: "851d4e8d-09ef-4acf-b813-c3b715ff54b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.818993 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.819031 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:26 crc kubenswrapper[4799]: I0930 14:39:26.819042 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/851d4e8d-09ef-4acf-b813-c3b715ff54b2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.066530 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85c8f46c9-sxlj2"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.094467 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f99689854-glm7t"] Sep 30 14:39:27 crc kubenswrapper[4799]: E0930 14:39:27.097587 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerName="glance-log" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.097797 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerName="glance-log" Sep 30 14:39:27 crc kubenswrapper[4799]: E0930 14:39:27.097939 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerName="glance-httpd" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.098020 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerName="glance-httpd" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.109840 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerName="glance-httpd" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.110216 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" containerName="glance-log" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.113635 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f99689854-glm7t"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.113811 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.112924 4799 scope.go:117] "RemoveContainer" containerID="57c7031a2996e8bdded23e20b9ae624ead1c3c9db9e31352b6b09c69a5b5f18d" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.143967 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.216240 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77997c977-fsfdt"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.238488 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15890bb-d088-4126-8382-b5d8e423b608-logs\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.238732 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-tls-certs\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.238818 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-combined-ca-bundle\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.238846 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-config-data\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.238875 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-secret-key\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.258351 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-scripts\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.258415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8ggr\" (UniqueName: \"kubernetes.io/projected/d15890bb-d088-4126-8382-b5d8e423b608-kube-api-access-b8ggr\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.268858 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.278084 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f7c888686-pwv2q"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.280447 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.291129 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.296045 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f7c888686-pwv2q"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.361251 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-config-data\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.362082 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-logs\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.362393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-scripts\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.362549 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkpbj\" (UniqueName: \"kubernetes.io/projected/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-kube-api-access-bkpbj\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.362253 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.378872 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.390372 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15890bb-d088-4126-8382-b5d8e423b608-logs\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.391287 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-horizon-tls-certs\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.391718 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.390938 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15890bb-d088-4126-8382-b5d8e423b608-logs\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.391717 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.390968 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.393152 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-tls-certs\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.393850 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-combined-ca-bundle\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.394027 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-combined-ca-bundle\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.410197 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-config-data\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.410562 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-secret-key\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.410779 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-scripts\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.420333 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-scripts\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.436991 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-secret-key\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.432484 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-tls-certs\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.438574 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-combined-ca-bundle\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.438814 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-config-data\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.440506 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8ggr\" (UniqueName: \"kubernetes.io/projected/d15890bb-d088-4126-8382-b5d8e423b608-kube-api-access-b8ggr\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.441627 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-horizon-secret-key\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.491844 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8ggr\" (UniqueName: \"kubernetes.io/projected/d15890bb-d088-4126-8382-b5d8e423b608-kube-api-access-b8ggr\") pod \"horizon-7f99689854-glm7t\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.527714 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.546679 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-scripts\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.547053 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-config-data\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.547226 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkpbj\" (UniqueName: \"kubernetes.io/projected/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-kube-api-access-bkpbj\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.547347 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-scripts\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.547512 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.547660 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.547875 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-horizon-tls-certs\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.548048 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-combined-ca-bundle\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.548200 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.548344 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c542m\" (UniqueName: \"kubernetes.io/projected/263ea967-0679-4f9f-81f3-4f1eb6b20835-kube-api-access-c542m\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.548570 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.548734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-horizon-secret-key\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.549007 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-logs\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.549164 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-config-data\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.549313 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-logs\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.556169 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-logs\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.562392 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-scripts\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.569225 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-config-data\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.574705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-horizon-secret-key\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.574885 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-combined-ca-bundle\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.585673 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-horizon-tls-certs\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.609356 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkpbj\" (UniqueName: \"kubernetes.io/projected/bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6-kube-api-access-bkpbj\") pod \"horizon-7f7c888686-pwv2q\" (UID: \"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6\") " pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.651491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.651983 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c542m\" (UniqueName: \"kubernetes.io/projected/263ea967-0679-4f9f-81f3-4f1eb6b20835-kube-api-access-c542m\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.652059 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.652160 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-logs\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.652290 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-config-data\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.652315 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-scripts\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.652357 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.652380 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.652638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.653149 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-logs\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.657756 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.680383 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.680812 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.682733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-scripts\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.699416 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.700414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-config-data\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.719745 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c542m\" (UniqueName: \"kubernetes.io/projected/263ea967-0679-4f9f-81f3-4f1eb6b20835-kube-api-access-c542m\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.857024 4799 generic.go:334] "Generic (PLEG): container finished" podID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerID="2561537a6d88c97f0e979a2ca7be6ea1e4f84566d5aabd02607c41b15c723eb3" exitCode=143 Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.857095 4799 generic.go:334] "Generic (PLEG): container finished" podID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerID="623dcd6a3e46baf1a0ff371edd5775b5e1092f41a9a55d17fa8b660b1166c953" exitCode=143 Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.857136 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67d7038d-99dc-47b8-9a91-67df94bcb35f","Type":"ContainerDied","Data":"2561537a6d88c97f0e979a2ca7be6ea1e4f84566d5aabd02607c41b15c723eb3"} Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.857187 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67d7038d-99dc-47b8-9a91-67df94bcb35f","Type":"ContainerDied","Data":"623dcd6a3e46baf1a0ff371edd5775b5e1092f41a9a55d17fa8b660b1166c953"} Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.892262 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " pod="openstack/glance-default-external-api-0" Sep 30 14:39:27 crc kubenswrapper[4799]: I0930 14:39:27.923428 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.138190 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.274929 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-scripts\") pod \"67d7038d-99dc-47b8-9a91-67df94bcb35f\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.275048 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-combined-ca-bundle\") pod \"67d7038d-99dc-47b8-9a91-67df94bcb35f\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.275093 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"67d7038d-99dc-47b8-9a91-67df94bcb35f\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.275143 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-internal-tls-certs\") pod \"67d7038d-99dc-47b8-9a91-67df94bcb35f\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.275219 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-config-data\") pod \"67d7038d-99dc-47b8-9a91-67df94bcb35f\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.275308 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csj2f\" (UniqueName: \"kubernetes.io/projected/67d7038d-99dc-47b8-9a91-67df94bcb35f-kube-api-access-csj2f\") pod \"67d7038d-99dc-47b8-9a91-67df94bcb35f\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.275360 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-httpd-run\") pod \"67d7038d-99dc-47b8-9a91-67df94bcb35f\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.275394 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-logs\") pod \"67d7038d-99dc-47b8-9a91-67df94bcb35f\" (UID: \"67d7038d-99dc-47b8-9a91-67df94bcb35f\") " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.276503 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-logs" (OuterVolumeSpecName: "logs") pod "67d7038d-99dc-47b8-9a91-67df94bcb35f" (UID: "67d7038d-99dc-47b8-9a91-67df94bcb35f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.280038 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "67d7038d-99dc-47b8-9a91-67df94bcb35f" (UID: "67d7038d-99dc-47b8-9a91-67df94bcb35f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.309928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d7038d-99dc-47b8-9a91-67df94bcb35f-kube-api-access-csj2f" (OuterVolumeSpecName: "kube-api-access-csj2f") pod "67d7038d-99dc-47b8-9a91-67df94bcb35f" (UID: "67d7038d-99dc-47b8-9a91-67df94bcb35f"). InnerVolumeSpecName "kube-api-access-csj2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.333719 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-scripts" (OuterVolumeSpecName: "scripts") pod "67d7038d-99dc-47b8-9a91-67df94bcb35f" (UID: "67d7038d-99dc-47b8-9a91-67df94bcb35f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.365544 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "67d7038d-99dc-47b8-9a91-67df94bcb35f" (UID: "67d7038d-99dc-47b8-9a91-67df94bcb35f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.379073 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.379140 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.379155 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csj2f\" (UniqueName: \"kubernetes.io/projected/67d7038d-99dc-47b8-9a91-67df94bcb35f-kube-api-access-csj2f\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.379171 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.379185 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67d7038d-99dc-47b8-9a91-67df94bcb35f-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.492078 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67d7038d-99dc-47b8-9a91-67df94bcb35f" (UID: "67d7038d-99dc-47b8-9a91-67df94bcb35f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.581024 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-config-data" (OuterVolumeSpecName: "config-data") pod "67d7038d-99dc-47b8-9a91-67df94bcb35f" (UID: "67d7038d-99dc-47b8-9a91-67df94bcb35f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.590522 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="851d4e8d-09ef-4acf-b813-c3b715ff54b2" path="/var/lib/kubelet/pods/851d4e8d-09ef-4acf-b813-c3b715ff54b2/volumes" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.630223 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.630255 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.767859 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "67d7038d-99dc-47b8-9a91-67df94bcb35f" (UID: "67d7038d-99dc-47b8-9a91-67df94bcb35f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.848472 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67d7038d-99dc-47b8-9a91-67df94bcb35f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.854350 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.921181 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67d7038d-99dc-47b8-9a91-67df94bcb35f","Type":"ContainerDied","Data":"128746b890dbcd009cc49b08b31e01a51c0871cdb8fa38307e310f6ca34f0b20"} Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.921268 4799 scope.go:117] "RemoveContainer" containerID="2561537a6d88c97f0e979a2ca7be6ea1e4f84566d5aabd02607c41b15c723eb3" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.921604 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:28 crc kubenswrapper[4799]: I0930 14:39:28.951479 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.076513 4799 scope.go:117] "RemoveContainer" containerID="623dcd6a3e46baf1a0ff371edd5775b5e1092f41a9a55d17fa8b660b1166c953" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.084861 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.106723 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.138207 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:29 crc kubenswrapper[4799]: E0930 14:39:29.139687 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerName="glance-log" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.139714 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerName="glance-log" Sep 30 14:39:29 crc kubenswrapper[4799]: E0930 14:39:29.139769 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerName="glance-httpd" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.139792 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerName="glance-httpd" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.140105 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerName="glance-log" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.140128 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" containerName="glance-httpd" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.142378 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.146642 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.147761 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.194362 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f7c888686-pwv2q"] Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.220010 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.264729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.264850 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.264886 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.264997 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.265056 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.265124 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.265173 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnj2r\" (UniqueName: \"kubernetes.io/projected/27e3a51e-4278-4183-8329-15916d7304a7-kube-api-access-vnj2r\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.265320 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.328225 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f99689854-glm7t"] Sep 30 14:39:29 crc kubenswrapper[4799]: W0930 14:39:29.355498 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd15890bb_d088_4126_8382_b5d8e423b608.slice/crio-65cfd9ee288b6b3d104f178a313a71390fc9ae06d91f2a0f13adf2190c01ea70 WatchSource:0}: Error finding container 65cfd9ee288b6b3d104f178a313a71390fc9ae06d91f2a0f13adf2190c01ea70: Status 404 returned error can't find the container with id 65cfd9ee288b6b3d104f178a313a71390fc9ae06d91f2a0f13adf2190c01ea70 Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.376297 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.376406 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.376446 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.376490 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.376524 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnj2r\" (UniqueName: \"kubernetes.io/projected/27e3a51e-4278-4183-8329-15916d7304a7-kube-api-access-vnj2r\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.376602 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.376637 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.376699 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.377197 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.382142 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.391135 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.393335 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.394563 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.419292 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.431034 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.439638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnj2r\" (UniqueName: \"kubernetes.io/projected/27e3a51e-4278-4183-8329-15916d7304a7-kube-api-access-vnj2r\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.491857 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.706596 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:39:29 crc kubenswrapper[4799]: W0930 14:39:29.727694 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod263ea967_0679_4f9f_81f3_4f1eb6b20835.slice/crio-0ae4407a355479943f2fa9625694789170d4eeb69fc4ae873f199fab6152f230 WatchSource:0}: Error finding container 0ae4407a355479943f2fa9625694789170d4eeb69fc4ae873f199fab6152f230: Status 404 returned error can't find the container with id 0ae4407a355479943f2fa9625694789170d4eeb69fc4ae873f199fab6152f230 Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.774115 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.976535 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerStarted","Data":"a4fbd4432153b13e314e42c57c2f31fa2e9ece1a2b8200e3b7313f48f6109a34"} Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.980537 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"263ea967-0679-4f9f-81f3-4f1eb6b20835","Type":"ContainerStarted","Data":"0ae4407a355479943f2fa9625694789170d4eeb69fc4ae873f199fab6152f230"} Sep 30 14:39:29 crc kubenswrapper[4799]: I0930 14:39:29.983445 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerStarted","Data":"65cfd9ee288b6b3d104f178a313a71390fc9ae06d91f2a0f13adf2190c01ea70"} Sep 30 14:39:30 crc kubenswrapper[4799]: I0930 14:39:30.567227 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d7038d-99dc-47b8-9a91-67df94bcb35f" path="/var/lib/kubelet/pods/67d7038d-99dc-47b8-9a91-67df94bcb35f/volumes" Sep 30 14:39:30 crc kubenswrapper[4799]: I0930 14:39:30.659264 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:39:31 crc kubenswrapper[4799]: I0930 14:39:31.433996 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:39:31 crc kubenswrapper[4799]: I0930 14:39:31.534863 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6dhwn"] Sep 30 14:39:31 crc kubenswrapper[4799]: I0930 14:39:31.535528 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="dnsmasq-dns" containerID="cri-o://a8e149f9114058145cb9cc5f486f2cff9cd522f651cd10b8d2b4cd814d21d9a1" gracePeriod=10 Sep 30 14:39:32 crc kubenswrapper[4799]: I0930 14:39:32.012517 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"263ea967-0679-4f9f-81f3-4f1eb6b20835","Type":"ContainerStarted","Data":"6fe6ceb0e8172a311716a0c0048b38d77e10ec63ad8a357eabc21d83f5f7e320"} Sep 30 14:39:32 crc kubenswrapper[4799]: I0930 14:39:32.017221 4799 generic.go:334] "Generic (PLEG): container finished" podID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerID="a8e149f9114058145cb9cc5f486f2cff9cd522f651cd10b8d2b4cd814d21d9a1" exitCode=0 Sep 30 14:39:32 crc kubenswrapper[4799]: I0930 14:39:32.017521 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" event={"ID":"a98d7efa-190b-4499-b6ba-85aadf66ecae","Type":"ContainerDied","Data":"a8e149f9114058145cb9cc5f486f2cff9cd522f651cd10b8d2b4cd814d21d9a1"} Sep 30 14:39:32 crc kubenswrapper[4799]: I0930 14:39:32.741315 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Sep 30 14:39:33 crc kubenswrapper[4799]: I0930 14:39:33.039088 4799 generic.go:334] "Generic (PLEG): container finished" podID="ace1f710-ccd4-4825-8a6d-b66d508a217b" containerID="f5239377f75f41afe5178e8987a478a2f4461fff97d01826fa1359713ddf2fce" exitCode=0 Sep 30 14:39:33 crc kubenswrapper[4799]: I0930 14:39:33.039166 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6zlj6" event={"ID":"ace1f710-ccd4-4825-8a6d-b66d508a217b","Type":"ContainerDied","Data":"f5239377f75f41afe5178e8987a478a2f4461fff97d01826fa1359713ddf2fce"} Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.290456 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.417023 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqxn4\" (UniqueName: \"kubernetes.io/projected/ace1f710-ccd4-4825-8a6d-b66d508a217b-kube-api-access-rqxn4\") pod \"ace1f710-ccd4-4825-8a6d-b66d508a217b\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.417323 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-scripts\") pod \"ace1f710-ccd4-4825-8a6d-b66d508a217b\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.417449 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-fernet-keys\") pod \"ace1f710-ccd4-4825-8a6d-b66d508a217b\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.417485 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-combined-ca-bundle\") pod \"ace1f710-ccd4-4825-8a6d-b66d508a217b\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.417539 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-credential-keys\") pod \"ace1f710-ccd4-4825-8a6d-b66d508a217b\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.417556 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-config-data\") pod \"ace1f710-ccd4-4825-8a6d-b66d508a217b\" (UID: \"ace1f710-ccd4-4825-8a6d-b66d508a217b\") " Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.422968 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ace1f710-ccd4-4825-8a6d-b66d508a217b" (UID: "ace1f710-ccd4-4825-8a6d-b66d508a217b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.424806 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ace1f710-ccd4-4825-8a6d-b66d508a217b-kube-api-access-rqxn4" (OuterVolumeSpecName: "kube-api-access-rqxn4") pod "ace1f710-ccd4-4825-8a6d-b66d508a217b" (UID: "ace1f710-ccd4-4825-8a6d-b66d508a217b"). InnerVolumeSpecName "kube-api-access-rqxn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.429954 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-scripts" (OuterVolumeSpecName: "scripts") pod "ace1f710-ccd4-4825-8a6d-b66d508a217b" (UID: "ace1f710-ccd4-4825-8a6d-b66d508a217b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.438792 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ace1f710-ccd4-4825-8a6d-b66d508a217b" (UID: "ace1f710-ccd4-4825-8a6d-b66d508a217b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.451224 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ace1f710-ccd4-4825-8a6d-b66d508a217b" (UID: "ace1f710-ccd4-4825-8a6d-b66d508a217b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.452992 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-config-data" (OuterVolumeSpecName: "config-data") pod "ace1f710-ccd4-4825-8a6d-b66d508a217b" (UID: "ace1f710-ccd4-4825-8a6d-b66d508a217b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.521440 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.521487 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.521505 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqxn4\" (UniqueName: \"kubernetes.io/projected/ace1f710-ccd4-4825-8a6d-b66d508a217b-kube-api-access-rqxn4\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.521522 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.521536 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:36 crc kubenswrapper[4799]: I0930 14:39:36.521548 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ace1f710-ccd4-4825-8a6d-b66d508a217b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.108707 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6zlj6" event={"ID":"ace1f710-ccd4-4825-8a6d-b66d508a217b","Type":"ContainerDied","Data":"f3c68016490a25dd53bd16f1573912f6921288ae36e5a7e6fb85b6fb0af37ad8"} Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.108800 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3c68016490a25dd53bd16f1573912f6921288ae36e5a7e6fb85b6fb0af37ad8" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.108798 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6zlj6" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.405576 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6zlj6"] Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.417786 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6zlj6"] Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.499951 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-l7xw2"] Sep 30 14:39:37 crc kubenswrapper[4799]: E0930 14:39:37.505322 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace1f710-ccd4-4825-8a6d-b66d508a217b" containerName="keystone-bootstrap" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.505377 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace1f710-ccd4-4825-8a6d-b66d508a217b" containerName="keystone-bootstrap" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.505781 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ace1f710-ccd4-4825-8a6d-b66d508a217b" containerName="keystone-bootstrap" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.506892 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.514029 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.514077 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.514130 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.514412 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vnkdb" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.529430 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-l7xw2"] Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.649067 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-credential-keys\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.649255 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-config-data\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.649282 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-scripts\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.649451 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czsll\" (UniqueName: \"kubernetes.io/projected/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-kube-api-access-czsll\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.649674 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-fernet-keys\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.649858 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-combined-ca-bundle\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.740881 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.752230 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-config-data\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.752308 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-scripts\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.752354 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czsll\" (UniqueName: \"kubernetes.io/projected/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-kube-api-access-czsll\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.752411 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-fernet-keys\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.752482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-combined-ca-bundle\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.752521 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-credential-keys\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.758519 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-credential-keys\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.758637 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-fernet-keys\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.760010 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-scripts\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.766477 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-combined-ca-bundle\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.769086 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-config-data\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.773291 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czsll\" (UniqueName: \"kubernetes.io/projected/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-kube-api-access-czsll\") pod \"keystone-bootstrap-l7xw2\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:37 crc kubenswrapper[4799]: I0930 14:39:37.841556 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:39:38 crc kubenswrapper[4799]: I0930 14:39:38.518071 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ace1f710-ccd4-4825-8a6d-b66d508a217b" path="/var/lib/kubelet/pods/ace1f710-ccd4-4825-8a6d-b66d508a217b/volumes" Sep 30 14:39:42 crc kubenswrapper[4799]: W0930 14:39:42.593790 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27e3a51e_4278_4183_8329_15916d7304a7.slice/crio-be3e4333524dfa827545ef0ce8b5f69d33fcf536316bfa9742f9e4aaf44edc8d WatchSource:0}: Error finding container be3e4333524dfa827545ef0ce8b5f69d33fcf536316bfa9742f9e4aaf44edc8d: Status 404 returned error can't find the container with id be3e4333524dfa827545ef0ce8b5f69d33fcf536316bfa9742f9e4aaf44edc8d Sep 30 14:39:42 crc kubenswrapper[4799]: I0930 14:39:42.740843 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Sep 30 14:39:42 crc kubenswrapper[4799]: I0930 14:39:42.741335 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:39:43 crc kubenswrapper[4799]: I0930 14:39:43.178914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"27e3a51e-4278-4183-8329-15916d7304a7","Type":"ContainerStarted","Data":"be3e4333524dfa827545ef0ce8b5f69d33fcf536316bfa9742f9e4aaf44edc8d"} Sep 30 14:39:44 crc kubenswrapper[4799]: E0930 14:39:44.663774 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Sep 30 14:39:44 crc kubenswrapper[4799]: E0930 14:39:44.664073 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n649hbdh559h58ch66dhcfh657h77h574h5f5h5c4h8fh598h5cch586h86h679h689h67dh649h578h5c7h548h64bh699h5h78h587hd5h59bh9fhb5q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bkpbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7f7c888686-pwv2q_openstack(bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:39:44 crc kubenswrapper[4799]: E0930 14:39:44.667227 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" Sep 30 14:39:45 crc kubenswrapper[4799]: E0930 14:39:45.207262 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" Sep 30 14:39:45 crc kubenswrapper[4799]: E0930 14:39:45.900325 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Sep 30 14:39:45 crc kubenswrapper[4799]: E0930 14:39:45.900590 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nddh5d8h647h74h59h57h647h5b7h549h558h5b8h67bh659h54dh5bh5bfh7dh74h578h5dbh68dh565h645h5bh64bh5b6h54dh576h596h5d5h65bh649q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5rch4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-576cc8ddf-zr27w_openstack(c78ce9ff-e63b-4587-8943-daf6fe7b4b82): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:39:45 crc kubenswrapper[4799]: E0930 14:39:45.927979 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-576cc8ddf-zr27w" podUID="c78ce9ff-e63b-4587-8943-daf6fe7b4b82" Sep 30 14:39:45 crc kubenswrapper[4799]: E0930 14:39:45.958924 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Sep 30 14:39:45 crc kubenswrapper[4799]: E0930 14:39:45.959173 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n676hf6h59fh76h8dh647hbdh55dh574h5fh565h5ch5chdbh565hcdh5ch576h545h555h64dh58bhcfh55h555h5d6h594h5f9h8fh8ch5fh554q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wx7zv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-85c8f46c9-sxlj2_openstack(cd86e414-219a-4754-85c4-239ef97843c1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:39:45 crc kubenswrapper[4799]: E0930 14:39:45.964624 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-85c8f46c9-sxlj2" podUID="cd86e414-219a-4754-85c4-239ef97843c1" Sep 30 14:39:47 crc kubenswrapper[4799]: E0930 14:39:47.932065 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Sep 30 14:39:47 crc kubenswrapper[4799]: E0930 14:39:47.933439 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n66fhc6h66bh56dh64h687h578h8dhc4h576h5d7hcbh675h57ch99h5d8h9bh6dh78h5c7hd6h77h76hf8h59fh589h57dh5f9h66dh57dhd6h5cbq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jcknk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-77997c977-fsfdt_openstack(37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:39:47 crc kubenswrapper[4799]: E0930 14:39:47.936828 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-77997c977-fsfdt" podUID="37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.025223 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.178204 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-sb\") pod \"a98d7efa-190b-4499-b6ba-85aadf66ecae\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.178868 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzj9w\" (UniqueName: \"kubernetes.io/projected/a98d7efa-190b-4499-b6ba-85aadf66ecae-kube-api-access-wzj9w\") pod \"a98d7efa-190b-4499-b6ba-85aadf66ecae\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.179004 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-swift-storage-0\") pod \"a98d7efa-190b-4499-b6ba-85aadf66ecae\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.179035 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-nb\") pod \"a98d7efa-190b-4499-b6ba-85aadf66ecae\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.179071 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-svc\") pod \"a98d7efa-190b-4499-b6ba-85aadf66ecae\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.179100 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config\") pod \"a98d7efa-190b-4499-b6ba-85aadf66ecae\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.188142 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a98d7efa-190b-4499-b6ba-85aadf66ecae-kube-api-access-wzj9w" (OuterVolumeSpecName: "kube-api-access-wzj9w") pod "a98d7efa-190b-4499-b6ba-85aadf66ecae" (UID: "a98d7efa-190b-4499-b6ba-85aadf66ecae"). InnerVolumeSpecName "kube-api-access-wzj9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.242454 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a98d7efa-190b-4499-b6ba-85aadf66ecae" (UID: "a98d7efa-190b-4499-b6ba-85aadf66ecae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.247659 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.247723 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" event={"ID":"a98d7efa-190b-4499-b6ba-85aadf66ecae","Type":"ContainerDied","Data":"e03c857718c12876b33b8d6f3bb8d857c024b688d14d81495c6dce6fde665c94"} Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.247786 4799 scope.go:117] "RemoveContainer" containerID="a8e149f9114058145cb9cc5f486f2cff9cd522f651cd10b8d2b4cd814d21d9a1" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.255222 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a98d7efa-190b-4499-b6ba-85aadf66ecae" (UID: "a98d7efa-190b-4499-b6ba-85aadf66ecae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.284851 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.284901 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.284920 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzj9w\" (UniqueName: \"kubernetes.io/projected/a98d7efa-190b-4499-b6ba-85aadf66ecae-kube-api-access-wzj9w\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:48 crc kubenswrapper[4799]: E0930 14:39:48.293384 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config podName:a98d7efa-190b-4499-b6ba-85aadf66ecae nodeName:}" failed. No retries permitted until 2025-09-30 14:39:48.793347671 +0000 UTC m=+1210.876948098 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config") pod "a98d7efa-190b-4499-b6ba-85aadf66ecae" (UID: "a98d7efa-190b-4499-b6ba-85aadf66ecae") : error deleting /var/lib/kubelet/pods/a98d7efa-190b-4499-b6ba-85aadf66ecae/volume-subpaths: remove /var/lib/kubelet/pods/a98d7efa-190b-4499-b6ba-85aadf66ecae/volume-subpaths: no such file or directory Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.293638 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a98d7efa-190b-4499-b6ba-85aadf66ecae" (UID: "a98d7efa-190b-4499-b6ba-85aadf66ecae"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.293938 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a98d7efa-190b-4499-b6ba-85aadf66ecae" (UID: "a98d7efa-190b-4499-b6ba-85aadf66ecae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.387381 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.387429 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.793872 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config\") pod \"a98d7efa-190b-4499-b6ba-85aadf66ecae\" (UID: \"a98d7efa-190b-4499-b6ba-85aadf66ecae\") " Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.794945 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config" (OuterVolumeSpecName: "config") pod "a98d7efa-190b-4499-b6ba-85aadf66ecae" (UID: "a98d7efa-190b-4499-b6ba-85aadf66ecae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.891029 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6dhwn"] Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.896055 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a98d7efa-190b-4499-b6ba-85aadf66ecae-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:48 crc kubenswrapper[4799]: I0930 14:39:48.900676 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-6dhwn"] Sep 30 14:39:50 crc kubenswrapper[4799]: I0930 14:39:50.515437 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" path="/var/lib/kubelet/pods/a98d7efa-190b-4499-b6ba-85aadf66ecae/volumes" Sep 30 14:39:52 crc kubenswrapper[4799]: I0930 14:39:52.741865 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-6dhwn" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Sep 30 14:39:58 crc kubenswrapper[4799]: E0930 14:39:58.498599 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Sep 30 14:39:58 crc kubenswrapper[4799]: E0930 14:39:58.499309 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n649h5d7hd8h86h5ddhf8h647h694h59ch65ch5fdh566h559h695h55h649h8fh655h696h58fhc8h68dh6h94hdh67dh59bh7h598h5bch5d5h565q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b8ggr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7f99689854-glm7t_openstack(d15890bb-d088-4126-8382-b5d8e423b608): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:39:58 crc kubenswrapper[4799]: E0930 14:39:58.510799 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Sep 30 14:39:58 crc kubenswrapper[4799]: E0930 14:39:58.511090 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tznlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-xt6j5_openstack(2d30b6ec-a2d1-4f93-932b-e78856253cc9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:39:58 crc kubenswrapper[4799]: E0930 14:39:58.512388 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-xt6j5" podUID="2d30b6ec-a2d1-4f93-932b-e78856253cc9" Sep 30 14:39:58 crc kubenswrapper[4799]: E0930 14:39:58.517471 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.640749 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.710331 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-scripts\") pod \"cd86e414-219a-4754-85c4-239ef97843c1\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.710412 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd86e414-219a-4754-85c4-239ef97843c1-horizon-secret-key\") pod \"cd86e414-219a-4754-85c4-239ef97843c1\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.710594 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx7zv\" (UniqueName: \"kubernetes.io/projected/cd86e414-219a-4754-85c4-239ef97843c1-kube-api-access-wx7zv\") pod \"cd86e414-219a-4754-85c4-239ef97843c1\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.710724 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd86e414-219a-4754-85c4-239ef97843c1-logs\") pod \"cd86e414-219a-4754-85c4-239ef97843c1\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.710805 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-config-data\") pod \"cd86e414-219a-4754-85c4-239ef97843c1\" (UID: \"cd86e414-219a-4754-85c4-239ef97843c1\") " Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.711139 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd86e414-219a-4754-85c4-239ef97843c1-logs" (OuterVolumeSpecName: "logs") pod "cd86e414-219a-4754-85c4-239ef97843c1" (UID: "cd86e414-219a-4754-85c4-239ef97843c1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.711508 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-config-data" (OuterVolumeSpecName: "config-data") pod "cd86e414-219a-4754-85c4-239ef97843c1" (UID: "cd86e414-219a-4754-85c4-239ef97843c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.711685 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd86e414-219a-4754-85c4-239ef97843c1-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.711707 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.711961 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-scripts" (OuterVolumeSpecName: "scripts") pod "cd86e414-219a-4754-85c4-239ef97843c1" (UID: "cd86e414-219a-4754-85c4-239ef97843c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.718420 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd86e414-219a-4754-85c4-239ef97843c1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "cd86e414-219a-4754-85c4-239ef97843c1" (UID: "cd86e414-219a-4754-85c4-239ef97843c1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.718584 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd86e414-219a-4754-85c4-239ef97843c1-kube-api-access-wx7zv" (OuterVolumeSpecName: "kube-api-access-wx7zv") pod "cd86e414-219a-4754-85c4-239ef97843c1" (UID: "cd86e414-219a-4754-85c4-239ef97843c1"). InnerVolumeSpecName "kube-api-access-wx7zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.813880 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd86e414-219a-4754-85c4-239ef97843c1-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.813930 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd86e414-219a-4754-85c4-239ef97843c1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:58 crc kubenswrapper[4799]: I0930 14:39:58.813945 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx7zv\" (UniqueName: \"kubernetes.io/projected/cd86e414-219a-4754-85c4-239ef97843c1-kube-api-access-wx7zv\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: E0930 14:39:59.045433 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Sep 30 14:39:59 crc kubenswrapper[4799]: E0930 14:39:59.045890 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n59fh654hc4h59fh569h97h556h55bhbbh78hdhbbh574h566h65h5d6hb6h685hf8h5c4h689h59ch544h67fh595h5c9h5cbh64dh5b6h57ch5fdh65q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f8nrj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(bc98dba3-94f3-4cc0-91b4-152a3fa006b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.112925 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.120381 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220237 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-scripts\") pod \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220279 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-config-data\") pod \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220351 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-config-data\") pod \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220393 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rch4\" (UniqueName: \"kubernetes.io/projected/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-kube-api-access-5rch4\") pod \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220422 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-scripts\") pod \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220456 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-logs\") pod \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220539 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-horizon-secret-key\") pod \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220621 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-horizon-secret-key\") pod \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\" (UID: \"c78ce9ff-e63b-4587-8943-daf6fe7b4b82\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220673 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-logs\") pod \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220758 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcknk\" (UniqueName: \"kubernetes.io/projected/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-kube-api-access-jcknk\") pod \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\" (UID: \"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8\") " Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.220835 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-scripts" (OuterVolumeSpecName: "scripts") pod "c78ce9ff-e63b-4587-8943-daf6fe7b4b82" (UID: "c78ce9ff-e63b-4587-8943-daf6fe7b4b82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.221150 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-logs" (OuterVolumeSpecName: "logs") pod "c78ce9ff-e63b-4587-8943-daf6fe7b4b82" (UID: "c78ce9ff-e63b-4587-8943-daf6fe7b4b82"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.221208 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.221607 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-config-data" (OuterVolumeSpecName: "config-data") pod "c78ce9ff-e63b-4587-8943-daf6fe7b4b82" (UID: "c78ce9ff-e63b-4587-8943-daf6fe7b4b82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.222495 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-logs" (OuterVolumeSpecName: "logs") pod "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8" (UID: "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.223985 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-scripts" (OuterVolumeSpecName: "scripts") pod "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8" (UID: "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.225384 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c78ce9ff-e63b-4587-8943-daf6fe7b4b82" (UID: "c78ce9ff-e63b-4587-8943-daf6fe7b4b82"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.225483 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-config-data" (OuterVolumeSpecName: "config-data") pod "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8" (UID: "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.225947 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-kube-api-access-jcknk" (OuterVolumeSpecName: "kube-api-access-jcknk") pod "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8" (UID: "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8"). InnerVolumeSpecName "kube-api-access-jcknk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.227497 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8" (UID: "37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.228087 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-kube-api-access-5rch4" (OuterVolumeSpecName: "kube-api-access-5rch4") pod "c78ce9ff-e63b-4587-8943-daf6fe7b4b82" (UID: "c78ce9ff-e63b-4587-8943-daf6fe7b4b82"). InnerVolumeSpecName "kube-api-access-5rch4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.329411 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.330042 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.330067 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcknk\" (UniqueName: \"kubernetes.io/projected/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-kube-api-access-jcknk\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.330083 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.330110 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.330121 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rch4\" (UniqueName: \"kubernetes.io/projected/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-kube-api-access-5rch4\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.330131 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.330143 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78ce9ff-e63b-4587-8943-daf6fe7b4b82-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.330152 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.432106 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85c8f46c9-sxlj2" event={"ID":"cd86e414-219a-4754-85c4-239ef97843c1","Type":"ContainerDied","Data":"8f4fcbbfb06ee03188e82cda7ec55ee9c4cab3bf2ea599fe8c926ad469c5ee2f"} Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.432208 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85c8f46c9-sxlj2" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.434804 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-576cc8ddf-zr27w" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.434794 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-576cc8ddf-zr27w" event={"ID":"c78ce9ff-e63b-4587-8943-daf6fe7b4b82","Type":"ContainerDied","Data":"9a5287604946212723bc5832648dcd12c158eb9db39e0235b4c9ebf29aec24e8"} Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.439303 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77997c977-fsfdt" event={"ID":"37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8","Type":"ContainerDied","Data":"1257a900b53f4751acbecd072d09c4012a5db5c27ee334003d8e84c5501df266"} Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.439398 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77997c977-fsfdt" Sep 30 14:39:59 crc kubenswrapper[4799]: E0930 14:39:59.444526 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-xt6j5" podUID="2d30b6ec-a2d1-4f93-932b-e78856253cc9" Sep 30 14:39:59 crc kubenswrapper[4799]: E0930 14:39:59.458353 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.607219 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-576cc8ddf-zr27w"] Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.635596 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-576cc8ddf-zr27w"] Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.651860 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85c8f46c9-sxlj2"] Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.664204 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85c8f46c9-sxlj2"] Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.696070 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77997c977-fsfdt"] Sep 30 14:39:59 crc kubenswrapper[4799]: I0930 14:39:59.707017 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-77997c977-fsfdt"] Sep 30 14:40:00 crc kubenswrapper[4799]: I0930 14:40:00.518437 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8" path="/var/lib/kubelet/pods/37a36ff7-5c03-45b6-ae06-b1c6dc2d10a8/volumes" Sep 30 14:40:00 crc kubenswrapper[4799]: I0930 14:40:00.519405 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c78ce9ff-e63b-4587-8943-daf6fe7b4b82" path="/var/lib/kubelet/pods/c78ce9ff-e63b-4587-8943-daf6fe7b4b82/volumes" Sep 30 14:40:00 crc kubenswrapper[4799]: I0930 14:40:00.520024 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd86e414-219a-4754-85c4-239ef97843c1" path="/var/lib/kubelet/pods/cd86e414-219a-4754-85c4-239ef97843c1/volumes" Sep 30 14:40:00 crc kubenswrapper[4799]: E0930 14:40:00.728797 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Sep 30 14:40:00 crc kubenswrapper[4799]: E0930 14:40:00.729056 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k2txv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-mx9fz_openstack(78ac655a-3fc7-46a4-93c2-69574e8b8f22): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 14:40:00 crc kubenswrapper[4799]: E0930 14:40:00.731016 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-mx9fz" podUID="78ac655a-3fc7-46a4-93c2-69574e8b8f22" Sep 30 14:40:00 crc kubenswrapper[4799]: I0930 14:40:00.747158 4799 scope.go:117] "RemoveContainer" containerID="febc20fa98082e3d1a188cbe33e9267d0a6f2de59c8f0288fa1227b75ebb2f05" Sep 30 14:40:01 crc kubenswrapper[4799]: I0930 14:40:01.259288 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-l7xw2"] Sep 30 14:40:01 crc kubenswrapper[4799]: W0930 14:40:01.280259 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c9ce36c_5f9d_43b2_8b0c_1cf52eb40229.slice/crio-68fbc3057aa9791c100c1da240a45da5a8f078a7de17df7302a4a759dd27bbaf WatchSource:0}: Error finding container 68fbc3057aa9791c100c1da240a45da5a8f078a7de17df7302a4a759dd27bbaf: Status 404 returned error can't find the container with id 68fbc3057aa9791c100c1da240a45da5a8f078a7de17df7302a4a759dd27bbaf Sep 30 14:40:01 crc kubenswrapper[4799]: I0930 14:40:01.468842 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerStarted","Data":"f4ee5dee783bb7f3fa7a55ca1a7dd48e0f8ecc369f9bdd149c1abc88def60447"} Sep 30 14:40:01 crc kubenswrapper[4799]: I0930 14:40:01.477184 4799 generic.go:334] "Generic (PLEG): container finished" podID="67e3be2e-5fd3-4ac7-9138-c968d3d0c025" containerID="85994786d1236ad433f7cb78913926fed48b7e24160914eb8d1d855e9768c479" exitCode=0 Sep 30 14:40:01 crc kubenswrapper[4799]: I0930 14:40:01.477280 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fb472" event={"ID":"67e3be2e-5fd3-4ac7-9138-c968d3d0c025","Type":"ContainerDied","Data":"85994786d1236ad433f7cb78913926fed48b7e24160914eb8d1d855e9768c479"} Sep 30 14:40:01 crc kubenswrapper[4799]: I0930 14:40:01.486924 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qrs4k" event={"ID":"e71e7667-1d26-47fc-9884-1c946e72c244","Type":"ContainerStarted","Data":"6f7d06a08493951cd016e08396484bd16fb6f4e1463d60160b07330decf52bfb"} Sep 30 14:40:01 crc kubenswrapper[4799]: I0930 14:40:01.489856 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l7xw2" event={"ID":"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229","Type":"ContainerStarted","Data":"68fbc3057aa9791c100c1da240a45da5a8f078a7de17df7302a4a759dd27bbaf"} Sep 30 14:40:01 crc kubenswrapper[4799]: E0930 14:40:01.490587 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-mx9fz" podUID="78ac655a-3fc7-46a4-93c2-69574e8b8f22" Sep 30 14:40:01 crc kubenswrapper[4799]: I0930 14:40:01.568423 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qrs4k" podStartSLOduration=5.81180832 podStartE2EDuration="46.568392903s" podCreationTimestamp="2025-09-30 14:39:15 +0000 UTC" firstStartedPulling="2025-09-30 14:39:18.284903238 +0000 UTC m=+1180.368503665" lastFinishedPulling="2025-09-30 14:39:59.041487821 +0000 UTC m=+1221.125088248" observedRunningTime="2025-09-30 14:40:01.554756674 +0000 UTC m=+1223.638357111" watchObservedRunningTime="2025-09-30 14:40:01.568392903 +0000 UTC m=+1223.651993330" Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.539667 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerStarted","Data":"0db7b3053e7c7468a8690c8b47ee3b620779c70dce53efef8b358e92bdd7e983"} Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.547238 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"263ea967-0679-4f9f-81f3-4f1eb6b20835","Type":"ContainerStarted","Data":"6cb6f4203908431a077c358319044c348ecce50af75f2791408d4e0f743a8a8a"} Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.550277 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc98dba3-94f3-4cc0-91b4-152a3fa006b7","Type":"ContainerStarted","Data":"1dcb7f3d5d8af2b3a16c5b094fa265e2b8dccf34c2420b99e5d60c425be29877"} Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.556915 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l7xw2" event={"ID":"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229","Type":"ContainerStarted","Data":"7f62f4f5059123bfcf58b12347dd2f0cb1506418ef5092f072080f309570c950"} Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.590193 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f7c888686-pwv2q" podStartSLOduration=3.993569093 podStartE2EDuration="35.590159993s" podCreationTimestamp="2025-09-30 14:39:27 +0000 UTC" firstStartedPulling="2025-09-30 14:39:29.225215347 +0000 UTC m=+1191.308815774" lastFinishedPulling="2025-09-30 14:40:00.821806237 +0000 UTC m=+1222.905406674" observedRunningTime="2025-09-30 14:40:02.562903516 +0000 UTC m=+1224.646503943" watchObservedRunningTime="2025-09-30 14:40:02.590159993 +0000 UTC m=+1224.673760420" Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.602858 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"27e3a51e-4278-4183-8329-15916d7304a7","Type":"ContainerStarted","Data":"ee708f9e443e77270b2fa71111914e5caec7831065bb86be29b62ea5f87db5e6"} Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.619768 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=35.619744467 podStartE2EDuration="35.619744467s" podCreationTimestamp="2025-09-30 14:39:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:02.602703019 +0000 UTC m=+1224.686303466" watchObservedRunningTime="2025-09-30 14:40:02.619744467 +0000 UTC m=+1224.703344894" Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.633730 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-l7xw2" podStartSLOduration=25.633706206 podStartE2EDuration="25.633706206s" podCreationTimestamp="2025-09-30 14:39:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:02.629008668 +0000 UTC m=+1224.712609095" watchObservedRunningTime="2025-09-30 14:40:02.633706206 +0000 UTC m=+1224.717306633" Sep 30 14:40:02 crc kubenswrapper[4799]: I0930 14:40:02.984120 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fb472" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.113457 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-config\") pod \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.113583 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-combined-ca-bundle\") pod \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.113730 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9459l\" (UniqueName: \"kubernetes.io/projected/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-kube-api-access-9459l\") pod \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\" (UID: \"67e3be2e-5fd3-4ac7-9138-c968d3d0c025\") " Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.126860 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-kube-api-access-9459l" (OuterVolumeSpecName: "kube-api-access-9459l") pod "67e3be2e-5fd3-4ac7-9138-c968d3d0c025" (UID: "67e3be2e-5fd3-4ac7-9138-c968d3d0c025"). InnerVolumeSpecName "kube-api-access-9459l". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.152863 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67e3be2e-5fd3-4ac7-9138-c968d3d0c025" (UID: "67e3be2e-5fd3-4ac7-9138-c968d3d0c025"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.167685 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-config" (OuterVolumeSpecName: "config") pod "67e3be2e-5fd3-4ac7-9138-c968d3d0c025" (UID: "67e3be2e-5fd3-4ac7-9138-c968d3d0c025"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.215824 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.215854 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9459l\" (UniqueName: \"kubernetes.io/projected/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-kube-api-access-9459l\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.215868 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/67e3be2e-5fd3-4ac7-9138-c968d3d0c025-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.618092 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fb472" event={"ID":"67e3be2e-5fd3-4ac7-9138-c968d3d0c025","Type":"ContainerDied","Data":"86b357cf5489ea254e95d900f8ca9196b26ab86b59ac53354f0840bb11a666f1"} Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.618451 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86b357cf5489ea254e95d900f8ca9196b26ab86b59ac53354f0840bb11a666f1" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.618124 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fb472" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.625111 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"27e3a51e-4278-4183-8329-15916d7304a7","Type":"ContainerStarted","Data":"d58da57a64dea1a04e5a5194dace7d97a3343dd15307cb9b3a2ca9c255f06efa"} Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.665349 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=34.665323814 podStartE2EDuration="34.665323814s" podCreationTimestamp="2025-09-30 14:39:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:03.66177257 +0000 UTC m=+1225.745373017" watchObservedRunningTime="2025-09-30 14:40:03.665323814 +0000 UTC m=+1225.748924241" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.876102 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-bzzsz"] Sep 30 14:40:03 crc kubenswrapper[4799]: E0930 14:40:03.876621 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="dnsmasq-dns" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.876670 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="dnsmasq-dns" Sep 30 14:40:03 crc kubenswrapper[4799]: E0930 14:40:03.876685 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e3be2e-5fd3-4ac7-9138-c968d3d0c025" containerName="neutron-db-sync" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.876692 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e3be2e-5fd3-4ac7-9138-c968d3d0c025" containerName="neutron-db-sync" Sep 30 14:40:03 crc kubenswrapper[4799]: E0930 14:40:03.876709 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="init" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.876716 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="init" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.876970 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a98d7efa-190b-4499-b6ba-85aadf66ecae" containerName="dnsmasq-dns" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.876989 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e3be2e-5fd3-4ac7-9138-c968d3d0c025" containerName="neutron-db-sync" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.878250 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.900894 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-bzzsz"] Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.947513 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnc8c\" (UniqueName: \"kubernetes.io/projected/afa0ef8b-835f-4d37-b610-3261288185db-kube-api-access-qnc8c\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.947629 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.947793 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.947816 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.947840 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:03 crc kubenswrapper[4799]: I0930 14:40:03.947870 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-config\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.044407 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-664b88d46b-sllrg"] Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.046533 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.049397 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.053303 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.053303 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tzfkp" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.053471 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.053871 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.059834 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.060191 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.060237 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.060279 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.060327 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-config\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.060424 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnc8c\" (UniqueName: \"kubernetes.io/projected/afa0ef8b-835f-4d37-b610-3261288185db-kube-api-access-qnc8c\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.061155 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.061643 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.061972 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.062457 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-config\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.083416 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-664b88d46b-sllrg"] Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.108229 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnc8c\" (UniqueName: \"kubernetes.io/projected/afa0ef8b-835f-4d37-b610-3261288185db-kube-api-access-qnc8c\") pod \"dnsmasq-dns-84b966f6c9-bzzsz\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.162811 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-config\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.164978 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-httpd-config\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.165105 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-ovndb-tls-certs\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.165171 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-combined-ca-bundle\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.165832 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5j57\" (UniqueName: \"kubernetes.io/projected/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-kube-api-access-z5j57\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.209724 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.272064 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-httpd-config\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.272393 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-ovndb-tls-certs\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.272421 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-combined-ca-bundle\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.272474 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5j57\" (UniqueName: \"kubernetes.io/projected/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-kube-api-access-z5j57\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.272561 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-config\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.290128 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-combined-ca-bundle\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.290615 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-config\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.296093 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-ovndb-tls-certs\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.304758 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-httpd-config\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.310111 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5j57\" (UniqueName: \"kubernetes.io/projected/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-kube-api-access-z5j57\") pod \"neutron-664b88d46b-sllrg\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.378055 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:04 crc kubenswrapper[4799]: I0930 14:40:04.938419 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-bzzsz"] Sep 30 14:40:05 crc kubenswrapper[4799]: I0930 14:40:05.344812 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-664b88d46b-sllrg"] Sep 30 14:40:05 crc kubenswrapper[4799]: I0930 14:40:05.674532 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-664b88d46b-sllrg" event={"ID":"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150","Type":"ContainerStarted","Data":"f6f84ddeb12c8e433a6cd607dcb07ac70d9dd3622d2231a3fe7019d9bd1e0f25"} Sep 30 14:40:05 crc kubenswrapper[4799]: I0930 14:40:05.686301 4799 generic.go:334] "Generic (PLEG): container finished" podID="afa0ef8b-835f-4d37-b610-3261288185db" containerID="46abd85e21d46e3e4bf3f246d44a5c1f83e78d0de2ae1f7e49b01ad61bb65083" exitCode=0 Sep 30 14:40:05 crc kubenswrapper[4799]: I0930 14:40:05.686337 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" event={"ID":"afa0ef8b-835f-4d37-b610-3261288185db","Type":"ContainerDied","Data":"46abd85e21d46e3e4bf3f246d44a5c1f83e78d0de2ae1f7e49b01ad61bb65083"} Sep 30 14:40:05 crc kubenswrapper[4799]: I0930 14:40:05.686437 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" event={"ID":"afa0ef8b-835f-4d37-b610-3261288185db","Type":"ContainerStarted","Data":"a77a3c40f8cd9d1f517ff1c987583ac19ac814ba64f60640b6ddb0499023b961"} Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.741330 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-664b88d46b-sllrg" event={"ID":"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150","Type":"ContainerStarted","Data":"d4173870f24110850c8b8a5559df233eebbb65a87542709008409ecb3e009c79"} Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.741774 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.741788 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-664b88d46b-sllrg" event={"ID":"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150","Type":"ContainerStarted","Data":"6609957b639734329472ddec0bca06f450dfeaaa8fb6132e950d36c48aea6e4a"} Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.748843 4799 generic.go:334] "Generic (PLEG): container finished" podID="e71e7667-1d26-47fc-9884-1c946e72c244" containerID="6f7d06a08493951cd016e08396484bd16fb6f4e1463d60160b07330decf52bfb" exitCode=0 Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.748938 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qrs4k" event={"ID":"e71e7667-1d26-47fc-9884-1c946e72c244","Type":"ContainerDied","Data":"6f7d06a08493951cd016e08396484bd16fb6f4e1463d60160b07330decf52bfb"} Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.755851 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" event={"ID":"afa0ef8b-835f-4d37-b610-3261288185db","Type":"ContainerStarted","Data":"e53d95b617c76885c77c6041547acd2f4243ccce407e260e09accbb601b17122"} Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.756018 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.765593 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-664b88d46b-sllrg" podStartSLOduration=3.765560186 podStartE2EDuration="3.765560186s" podCreationTimestamp="2025-09-30 14:40:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:06.758622053 +0000 UTC m=+1228.842222480" watchObservedRunningTime="2025-09-30 14:40:06.765560186 +0000 UTC m=+1228.849160613" Sep 30 14:40:06 crc kubenswrapper[4799]: I0930 14:40:06.817687 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" podStartSLOduration=3.817633148 podStartE2EDuration="3.817633148s" podCreationTimestamp="2025-09-30 14:40:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:06.81255784 +0000 UTC m=+1228.896158287" watchObservedRunningTime="2025-09-30 14:40:06.817633148 +0000 UTC m=+1228.901233575" Sep 30 14:40:07 crc kubenswrapper[4799]: I0930 14:40:07.681928 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:40:07 crc kubenswrapper[4799]: I0930 14:40:07.683572 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:40:07 crc kubenswrapper[4799]: I0930 14:40:07.926842 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 14:40:07 crc kubenswrapper[4799]: I0930 14:40:07.926919 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.000484 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.088050 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.259984 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-67c5b7d44f-24zhg"] Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.262473 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.285547 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.285701 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.306746 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67c5b7d44f-24zhg"] Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.405944 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-combined-ca-bundle\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.406018 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-public-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.406044 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-internal-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.406096 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-config\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.406128 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-ovndb-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.406155 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-httpd-config\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.406193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgd4v\" (UniqueName: \"kubernetes.io/projected/c013c008-ebe8-4afe-8a76-b9103ae99fa6-kube-api-access-lgd4v\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.507757 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-config\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.507844 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-ovndb-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.507884 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-httpd-config\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.507951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgd4v\" (UniqueName: \"kubernetes.io/projected/c013c008-ebe8-4afe-8a76-b9103ae99fa6-kube-api-access-lgd4v\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.508031 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-combined-ca-bundle\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.508071 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-public-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.508100 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-internal-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.517820 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-httpd-config\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.544526 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-public-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.544533 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-internal-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.546487 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-combined-ca-bundle\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.547296 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgd4v\" (UniqueName: \"kubernetes.io/projected/c013c008-ebe8-4afe-8a76-b9103ae99fa6-kube-api-access-lgd4v\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.547516 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-config\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.549643 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c013c008-ebe8-4afe-8a76-b9103ae99fa6-ovndb-tls-certs\") pod \"neutron-67c5b7d44f-24zhg\" (UID: \"c013c008-ebe8-4afe-8a76-b9103ae99fa6\") " pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.642338 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.774317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 14:40:08 crc kubenswrapper[4799]: I0930 14:40:08.774711 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 14:40:09 crc kubenswrapper[4799]: I0930 14:40:09.779752 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 14:40:09 crc kubenswrapper[4799]: I0930 14:40:09.780102 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 14:40:09 crc kubenswrapper[4799]: I0930 14:40:09.851242 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 14:40:09 crc kubenswrapper[4799]: I0930 14:40:09.872101 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 14:40:10 crc kubenswrapper[4799]: I0930 14:40:10.821985 4799 generic.go:334] "Generic (PLEG): container finished" podID="0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" containerID="7f62f4f5059123bfcf58b12347dd2f0cb1506418ef5092f072080f309570c950" exitCode=0 Sep 30 14:40:10 crc kubenswrapper[4799]: I0930 14:40:10.822451 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:40:10 crc kubenswrapper[4799]: I0930 14:40:10.822465 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:40:10 crc kubenswrapper[4799]: I0930 14:40:10.822865 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l7xw2" event={"ID":"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229","Type":"ContainerDied","Data":"7f62f4f5059123bfcf58b12347dd2f0cb1506418ef5092f072080f309570c950"} Sep 30 14:40:10 crc kubenswrapper[4799]: I0930 14:40:10.825277 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 14:40:10 crc kubenswrapper[4799]: I0930 14:40:10.825319 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.082310 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qrs4k" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.160056 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-config-data\") pod \"e71e7667-1d26-47fc-9884-1c946e72c244\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.160144 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71e7667-1d26-47fc-9884-1c946e72c244-logs\") pod \"e71e7667-1d26-47fc-9884-1c946e72c244\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.160268 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-scripts\") pod \"e71e7667-1d26-47fc-9884-1c946e72c244\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.160454 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpqzs\" (UniqueName: \"kubernetes.io/projected/e71e7667-1d26-47fc-9884-1c946e72c244-kube-api-access-rpqzs\") pod \"e71e7667-1d26-47fc-9884-1c946e72c244\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.160515 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-combined-ca-bundle\") pod \"e71e7667-1d26-47fc-9884-1c946e72c244\" (UID: \"e71e7667-1d26-47fc-9884-1c946e72c244\") " Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.161169 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e71e7667-1d26-47fc-9884-1c946e72c244-logs" (OuterVolumeSpecName: "logs") pod "e71e7667-1d26-47fc-9884-1c946e72c244" (UID: "e71e7667-1d26-47fc-9884-1c946e72c244"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.182379 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-scripts" (OuterVolumeSpecName: "scripts") pod "e71e7667-1d26-47fc-9884-1c946e72c244" (UID: "e71e7667-1d26-47fc-9884-1c946e72c244"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.204047 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e71e7667-1d26-47fc-9884-1c946e72c244-kube-api-access-rpqzs" (OuterVolumeSpecName: "kube-api-access-rpqzs") pod "e71e7667-1d26-47fc-9884-1c946e72c244" (UID: "e71e7667-1d26-47fc-9884-1c946e72c244"). InnerVolumeSpecName "kube-api-access-rpqzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.210220 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-config-data" (OuterVolumeSpecName: "config-data") pod "e71e7667-1d26-47fc-9884-1c946e72c244" (UID: "e71e7667-1d26-47fc-9884-1c946e72c244"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.252713 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e71e7667-1d26-47fc-9884-1c946e72c244" (UID: "e71e7667-1d26-47fc-9884-1c946e72c244"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.265134 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.265173 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpqzs\" (UniqueName: \"kubernetes.io/projected/e71e7667-1d26-47fc-9884-1c946e72c244-kube-api-access-rpqzs\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.265192 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.265202 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e71e7667-1d26-47fc-9884-1c946e72c244-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.265214 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e71e7667-1d26-47fc-9884-1c946e72c244-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.834409 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qrs4k" event={"ID":"e71e7667-1d26-47fc-9884-1c946e72c244","Type":"ContainerDied","Data":"4b7b993575c7d3ebbc1d07e25abb8a09e48284374ae9ad79f7c1c184672ede8a"} Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.835055 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b7b993575c7d3ebbc1d07e25abb8a09e48284374ae9ad79f7c1c184672ede8a" Sep 30 14:40:11 crc kubenswrapper[4799]: I0930 14:40:11.834462 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qrs4k" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.215699 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9fd6b988d-fckfn"] Sep 30 14:40:12 crc kubenswrapper[4799]: E0930 14:40:12.216189 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71e7667-1d26-47fc-9884-1c946e72c244" containerName="placement-db-sync" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.216214 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71e7667-1d26-47fc-9884-1c946e72c244" containerName="placement-db-sync" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.216430 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71e7667-1d26-47fc-9884-1c946e72c244" containerName="placement-db-sync" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.217459 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.225636 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.225916 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.226061 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.226276 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.231300 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-j4x5t" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.251501 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9fd6b988d-fckfn"] Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.291134 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79w6n\" (UniqueName: \"kubernetes.io/projected/23d1089b-5d01-44ec-b8de-617485c30c7a-kube-api-access-79w6n\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.291217 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-scripts\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.291253 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-combined-ca-bundle\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.291281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-public-tls-certs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.291331 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-config-data\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.291455 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-internal-tls-certs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.291482 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d1089b-5d01-44ec-b8de-617485c30c7a-logs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.392882 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-internal-tls-certs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.392958 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d1089b-5d01-44ec-b8de-617485c30c7a-logs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.393038 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79w6n\" (UniqueName: \"kubernetes.io/projected/23d1089b-5d01-44ec-b8de-617485c30c7a-kube-api-access-79w6n\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.393076 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-scripts\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.393109 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-combined-ca-bundle\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.393135 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-public-tls-certs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.393175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-config-data\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.398555 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d1089b-5d01-44ec-b8de-617485c30c7a-logs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.405975 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-combined-ca-bundle\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.410129 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-public-tls-certs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.410376 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-scripts\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.411077 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-internal-tls-certs\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.426445 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79w6n\" (UniqueName: \"kubernetes.io/projected/23d1089b-5d01-44ec-b8de-617485c30c7a-kube-api-access-79w6n\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.427046 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d1089b-5d01-44ec-b8de-617485c30c7a-config-data\") pod \"placement-9fd6b988d-fckfn\" (UID: \"23d1089b-5d01-44ec-b8de-617485c30c7a\") " pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.564280 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.849099 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:40:12 crc kubenswrapper[4799]: I0930 14:40:12.849130 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:40:14 crc kubenswrapper[4799]: I0930 14:40:14.211472 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:14 crc kubenswrapper[4799]: I0930 14:40:14.295602 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-mcsxn"] Sep 30 14:40:14 crc kubenswrapper[4799]: I0930 14:40:14.299555 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" podUID="89554c30-2812-4e26-9878-a972ab70dd7e" containerName="dnsmasq-dns" containerID="cri-o://fcfbbd41979a3563f6feebb2661150b16ce97a70f3f9e5bb53e5e85e4976b6d1" gracePeriod=10 Sep 30 14:40:14 crc kubenswrapper[4799]: I0930 14:40:14.871857 4799 generic.go:334] "Generic (PLEG): container finished" podID="89554c30-2812-4e26-9878-a972ab70dd7e" containerID="fcfbbd41979a3563f6feebb2661150b16ce97a70f3f9e5bb53e5e85e4976b6d1" exitCode=0 Sep 30 14:40:14 crc kubenswrapper[4799]: I0930 14:40:14.871968 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" event={"ID":"89554c30-2812-4e26-9878-a972ab70dd7e","Type":"ContainerDied","Data":"fcfbbd41979a3563f6feebb2661150b16ce97a70f3f9e5bb53e5e85e4976b6d1"} Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.902980 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.971039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l7xw2" event={"ID":"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229","Type":"ContainerDied","Data":"68fbc3057aa9791c100c1da240a45da5a8f078a7de17df7302a4a759dd27bbaf"} Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.971115 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68fbc3057aa9791c100c1da240a45da5a8f078a7de17df7302a4a759dd27bbaf" Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.971308 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l7xw2" Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.992480 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-combined-ca-bundle\") pod \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.992573 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-config-data\") pod \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.992610 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-scripts\") pod \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.992669 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-fernet-keys\") pod \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.992719 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czsll\" (UniqueName: \"kubernetes.io/projected/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-kube-api-access-czsll\") pod \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " Sep 30 14:40:15 crc kubenswrapper[4799]: I0930 14:40:15.992766 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-credential-keys\") pod \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\" (UID: \"0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229\") " Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.022267 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-kube-api-access-czsll" (OuterVolumeSpecName: "kube-api-access-czsll") pod "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" (UID: "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229"). InnerVolumeSpecName "kube-api-access-czsll". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.025809 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-scripts" (OuterVolumeSpecName: "scripts") pod "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" (UID: "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.040149 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" (UID: "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.040335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" (UID: "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.082858 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-config-data" (OuterVolumeSpecName: "config-data") pod "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" (UID: "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.133559 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" (UID: "0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.145137 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.145223 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.145295 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.145308 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.145319 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czsll\" (UniqueName: \"kubernetes.io/projected/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-kube-api-access-czsll\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.145332 4799 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229-credential-keys\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.311148 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.466169 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-svc\") pod \"89554c30-2812-4e26-9878-a972ab70dd7e\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.466262 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-nb\") pod \"89554c30-2812-4e26-9878-a972ab70dd7e\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.466349 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpxff\" (UniqueName: \"kubernetes.io/projected/89554c30-2812-4e26-9878-a972ab70dd7e-kube-api-access-cpxff\") pod \"89554c30-2812-4e26-9878-a972ab70dd7e\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.466405 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-config\") pod \"89554c30-2812-4e26-9878-a972ab70dd7e\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.466455 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-swift-storage-0\") pod \"89554c30-2812-4e26-9878-a972ab70dd7e\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.466517 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-sb\") pod \"89554c30-2812-4e26-9878-a972ab70dd7e\" (UID: \"89554c30-2812-4e26-9878-a972ab70dd7e\") " Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.494565 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89554c30-2812-4e26-9878-a972ab70dd7e-kube-api-access-cpxff" (OuterVolumeSpecName: "kube-api-access-cpxff") pod "89554c30-2812-4e26-9878-a972ab70dd7e" (UID: "89554c30-2812-4e26-9878-a972ab70dd7e"). InnerVolumeSpecName "kube-api-access-cpxff". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.559274 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9fd6b988d-fckfn"] Sep 30 14:40:16 crc kubenswrapper[4799]: W0930 14:40:16.567201 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23d1089b_5d01_44ec_b8de_617485c30c7a.slice/crio-839eccbacf4e2048fa49ede0c3ceb15c9d885462741febbd3228506c628f947d WatchSource:0}: Error finding container 839eccbacf4e2048fa49ede0c3ceb15c9d885462741febbd3228506c628f947d: Status 404 returned error can't find the container with id 839eccbacf4e2048fa49ede0c3ceb15c9d885462741febbd3228506c628f947d Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.572386 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpxff\" (UniqueName: \"kubernetes.io/projected/89554c30-2812-4e26-9878-a972ab70dd7e-kube-api-access-cpxff\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.777693 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "89554c30-2812-4e26-9878-a972ab70dd7e" (UID: "89554c30-2812-4e26-9878-a972ab70dd7e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.786878 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.806541 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "89554c30-2812-4e26-9878-a972ab70dd7e" (UID: "89554c30-2812-4e26-9878-a972ab70dd7e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.850577 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "89554c30-2812-4e26-9878-a972ab70dd7e" (UID: "89554c30-2812-4e26-9878-a972ab70dd7e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.895133 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.895179 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.917942 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "89554c30-2812-4e26-9878-a972ab70dd7e" (UID: "89554c30-2812-4e26-9878-a972ab70dd7e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.949987 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-67c5b7d44f-24zhg"] Sep 30 14:40:16 crc kubenswrapper[4799]: I0930 14:40:16.985987 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-config" (OuterVolumeSpecName: "config") pod "89554c30-2812-4e26-9878-a972ab70dd7e" (UID: "89554c30-2812-4e26-9878-a972ab70dd7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.002518 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.002552 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/89554c30-2812-4e26-9878-a972ab70dd7e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.034256 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9fd6b988d-fckfn" event={"ID":"23d1089b-5d01-44ec-b8de-617485c30c7a","Type":"ContainerStarted","Data":"839eccbacf4e2048fa49ede0c3ceb15c9d885462741febbd3228506c628f947d"} Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.041275 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" event={"ID":"89554c30-2812-4e26-9878-a972ab70dd7e","Type":"ContainerDied","Data":"c6e433ba9151efb327a6b5fcfdb18a5bdef822227e1b39e36e947be55b926178"} Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.041366 4799 scope.go:117] "RemoveContainer" containerID="fcfbbd41979a3563f6feebb2661150b16ce97a70f3f9e5bb53e5e85e4976b6d1" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.041568 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-mcsxn" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.058149 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xt6j5" event={"ID":"2d30b6ec-a2d1-4f93-932b-e78856253cc9","Type":"ContainerStarted","Data":"87839f98445eabda60d578fba0d8f5cbc36fdf4a86b85aaf163eaa53acb92000"} Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.111867 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc98dba3-94f3-4cc0-91b4-152a3fa006b7","Type":"ContainerStarted","Data":"7c8d2e81c5226bd8ee43d610fff5f34cde4c34313c054a0b2707edc9c2be2379"} Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.132497 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerStarted","Data":"8923b064d4910eddccd36258c0091f291d87f5d2b5ba8a2a2855d3bafc5f9665"} Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.134875 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-577d747d48-ldwh6"] Sep 30 14:40:17 crc kubenswrapper[4799]: E0930 14:40:17.135370 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89554c30-2812-4e26-9878-a972ab70dd7e" containerName="dnsmasq-dns" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.135393 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="89554c30-2812-4e26-9878-a972ab70dd7e" containerName="dnsmasq-dns" Sep 30 14:40:17 crc kubenswrapper[4799]: E0930 14:40:17.135410 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89554c30-2812-4e26-9878-a972ab70dd7e" containerName="init" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.135417 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="89554c30-2812-4e26-9878-a972ab70dd7e" containerName="init" Sep 30 14:40:17 crc kubenswrapper[4799]: E0930 14:40:17.135427 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" containerName="keystone-bootstrap" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.135433 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" containerName="keystone-bootstrap" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.135676 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="89554c30-2812-4e26-9878-a972ab70dd7e" containerName="dnsmasq-dns" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.135697 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" containerName="keystone-bootstrap" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.136570 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.137245 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xt6j5" podStartSLOduration=3.814688192 podStartE2EDuration="1m0.137221359s" podCreationTimestamp="2025-09-30 14:39:17 +0000 UTC" firstStartedPulling="2025-09-30 14:39:19.836683892 +0000 UTC m=+1181.920284309" lastFinishedPulling="2025-09-30 14:40:16.159217049 +0000 UTC m=+1238.242817476" observedRunningTime="2025-09-30 14:40:17.107038217 +0000 UTC m=+1239.190638654" watchObservedRunningTime="2025-09-30 14:40:17.137221359 +0000 UTC m=+1239.220821786" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.140746 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.150228 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.150508 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vnkdb" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.150545 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.150633 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.150834 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.197967 4799 scope.go:117] "RemoveContainer" containerID="99350693c10131040351e9357be558d8cd11740ed3bebb18ce0bf8d2bda72471" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.207497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-internal-tls-certs\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.207558 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qf7j\" (UniqueName: \"kubernetes.io/projected/301834e4-5636-4faa-b4b3-2b98f14791a7-kube-api-access-7qf7j\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.207700 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-config-data\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.207776 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-credential-keys\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.207818 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-fernet-keys\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.207853 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-combined-ca-bundle\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.207887 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-scripts\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.207921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-public-tls-certs\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.217737 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-577d747d48-ldwh6"] Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.253404 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-mcsxn"] Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.278026 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-mcsxn"] Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.309870 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-internal-tls-certs\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.309930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qf7j\" (UniqueName: \"kubernetes.io/projected/301834e4-5636-4faa-b4b3-2b98f14791a7-kube-api-access-7qf7j\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.310002 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-config-data\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.310032 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-credential-keys\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.310055 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-fernet-keys\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.310081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-combined-ca-bundle\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.310107 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-scripts\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.310132 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-public-tls-certs\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.320586 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-combined-ca-bundle\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.328218 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-scripts\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.329227 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-credential-keys\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.329611 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-public-tls-certs\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.330287 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-config-data\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.333318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-fernet-keys\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.341470 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/301834e4-5636-4faa-b4b3-2b98f14791a7-internal-tls-certs\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.352601 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qf7j\" (UniqueName: \"kubernetes.io/projected/301834e4-5636-4faa-b4b3-2b98f14791a7-kube-api-access-7qf7j\") pod \"keystone-577d747d48-ldwh6\" (UID: \"301834e4-5636-4faa-b4b3-2b98f14791a7\") " pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.487780 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:17 crc kubenswrapper[4799]: I0930 14:40:17.695267 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.186750 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9fd6b988d-fckfn" event={"ID":"23d1089b-5d01-44ec-b8de-617485c30c7a","Type":"ContainerStarted","Data":"404b0b05328f5a94403c8e7b9781c85a4d7a9b298389ceda9d3805eec6c71aba"} Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.186821 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9fd6b988d-fckfn" event={"ID":"23d1089b-5d01-44ec-b8de-617485c30c7a","Type":"ContainerStarted","Data":"37efccb2872bbe22d7dcd445969d71b51b61accac0d3db5a5d8d0d7a88d7f10d"} Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.188505 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.188566 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.229869 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerStarted","Data":"6218e563ed5620135720c08ff067c9d79266769a1728d09cab51ff7af21c5dfe"} Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.233161 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-mx9fz" event={"ID":"78ac655a-3fc7-46a4-93c2-69574e8b8f22","Type":"ContainerStarted","Data":"a86ae59bf3e2020b35f1021225e2b995c371d1ee708c374d3038fd248b6dcc86"} Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.246968 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c5b7d44f-24zhg" event={"ID":"c013c008-ebe8-4afe-8a76-b9103ae99fa6","Type":"ContainerStarted","Data":"5e1f78b6cc0ecb28a11564877bbc3d5c434387613284d8e356694246be878327"} Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.247026 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c5b7d44f-24zhg" event={"ID":"c013c008-ebe8-4afe-8a76-b9103ae99fa6","Type":"ContainerStarted","Data":"a764992d64e4f038e9018cc00712a2c4dc4ca7d36214c7e4039f4ac344ad53a3"} Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.259700 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-9fd6b988d-fckfn" podStartSLOduration=6.259671773 podStartE2EDuration="6.259671773s" podCreationTimestamp="2025-09-30 14:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:18.215060699 +0000 UTC m=+1240.298661126" watchObservedRunningTime="2025-09-30 14:40:18.259671773 +0000 UTC m=+1240.343272190" Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.260875 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f99689854-glm7t" podStartSLOduration=-9223371985.593908 podStartE2EDuration="51.260868198s" podCreationTimestamp="2025-09-30 14:39:27 +0000 UTC" firstStartedPulling="2025-09-30 14:39:29.412950975 +0000 UTC m=+1191.496551402" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:18.260457446 +0000 UTC m=+1240.344057873" watchObservedRunningTime="2025-09-30 14:40:18.260868198 +0000 UTC m=+1240.344468625" Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.288016 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-mx9fz" podStartSLOduration=4.461673197 podStartE2EDuration="1m1.287979731s" podCreationTimestamp="2025-09-30 14:39:17 +0000 UTC" firstStartedPulling="2025-09-30 14:39:19.326371554 +0000 UTC m=+1181.409971981" lastFinishedPulling="2025-09-30 14:40:16.152678088 +0000 UTC m=+1238.236278515" observedRunningTime="2025-09-30 14:40:18.28556882 +0000 UTC m=+1240.369169247" watchObservedRunningTime="2025-09-30 14:40:18.287979731 +0000 UTC m=+1240.371580158" Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.357717 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-577d747d48-ldwh6"] Sep 30 14:40:18 crc kubenswrapper[4799]: I0930 14:40:18.522757 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89554c30-2812-4e26-9878-a972ab70dd7e" path="/var/lib/kubelet/pods/89554c30-2812-4e26-9878-a972ab70dd7e/volumes" Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.289465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-577d747d48-ldwh6" event={"ID":"301834e4-5636-4faa-b4b3-2b98f14791a7","Type":"ContainerStarted","Data":"5d4ecc1855835371740ebede2de4bd145fb7b6df4c06c0c3058c30a502d3a940"} Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.289893 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-577d747d48-ldwh6" event={"ID":"301834e4-5636-4faa-b4b3-2b98f14791a7","Type":"ContainerStarted","Data":"beeb93133c4f1642cad0dd58116d1afd83a53a2c3c54867cfe64e0da27439a4d"} Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.290839 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.301738 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-67c5b7d44f-24zhg" event={"ID":"c013c008-ebe8-4afe-8a76-b9103ae99fa6","Type":"ContainerStarted","Data":"0de543b3844f498431f8cd1ff962c7ee9a681c9bb9c4dccfda53568518695270"} Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.326218 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-577d747d48-ldwh6" podStartSLOduration=2.326192672 podStartE2EDuration="2.326192672s" podCreationTimestamp="2025-09-30 14:40:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:19.321999009 +0000 UTC m=+1241.405599446" watchObservedRunningTime="2025-09-30 14:40:19.326192672 +0000 UTC m=+1241.409793109" Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.364032 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-67c5b7d44f-24zhg" podStartSLOduration=11.364001317 podStartE2EDuration="11.364001317s" podCreationTimestamp="2025-09-30 14:40:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:19.347049812 +0000 UTC m=+1241.430650239" watchObservedRunningTime="2025-09-30 14:40:19.364001317 +0000 UTC m=+1241.447601744" Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.374075 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.374197 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.501908 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.577167 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 14:40:19 crc kubenswrapper[4799]: I0930 14:40:19.577632 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 14:40:20 crc kubenswrapper[4799]: I0930 14:40:20.319169 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:25 crc kubenswrapper[4799]: I0930 14:40:25.400159 4799 generic.go:334] "Generic (PLEG): container finished" podID="2d30b6ec-a2d1-4f93-932b-e78856253cc9" containerID="87839f98445eabda60d578fba0d8f5cbc36fdf4a86b85aaf163eaa53acb92000" exitCode=0 Sep 30 14:40:25 crc kubenswrapper[4799]: I0930 14:40:25.400281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xt6j5" event={"ID":"2d30b6ec-a2d1-4f93-932b-e78856253cc9","Type":"ContainerDied","Data":"87839f98445eabda60d578fba0d8f5cbc36fdf4a86b85aaf163eaa53acb92000"} Sep 30 14:40:27 crc kubenswrapper[4799]: I0930 14:40:27.528334 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:40:27 crc kubenswrapper[4799]: I0930 14:40:27.531972 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:40:27 crc kubenswrapper[4799]: I0930 14:40:27.537550 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:40:27 crc kubenswrapper[4799]: I0930 14:40:27.682136 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.611106 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.696482 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-db-sync-config-data\") pod \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.697165 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-combined-ca-bundle\") pod \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.697232 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tznlh\" (UniqueName: \"kubernetes.io/projected/2d30b6ec-a2d1-4f93-932b-e78856253cc9-kube-api-access-tznlh\") pod \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\" (UID: \"2d30b6ec-a2d1-4f93-932b-e78856253cc9\") " Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.705129 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d30b6ec-a2d1-4f93-932b-e78856253cc9-kube-api-access-tznlh" (OuterVolumeSpecName: "kube-api-access-tznlh") pod "2d30b6ec-a2d1-4f93-932b-e78856253cc9" (UID: "2d30b6ec-a2d1-4f93-932b-e78856253cc9"). InnerVolumeSpecName "kube-api-access-tznlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.741824 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2d30b6ec-a2d1-4f93-932b-e78856253cc9" (UID: "2d30b6ec-a2d1-4f93-932b-e78856253cc9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.777061 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d30b6ec-a2d1-4f93-932b-e78856253cc9" (UID: "2d30b6ec-a2d1-4f93-932b-e78856253cc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.800346 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.800401 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d30b6ec-a2d1-4f93-932b-e78856253cc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:28 crc kubenswrapper[4799]: I0930 14:40:28.800411 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tznlh\" (UniqueName: \"kubernetes.io/projected/2d30b6ec-a2d1-4f93-932b-e78856253cc9-kube-api-access-tznlh\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.474580 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xt6j5" event={"ID":"2d30b6ec-a2d1-4f93-932b-e78856253cc9","Type":"ContainerDied","Data":"9b2ed069427d7da7cd9be63211e1d604a39fdf28d1b95686cb02474f506e43f4"} Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.474727 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b2ed069427d7da7cd9be63211e1d604a39fdf28d1b95686cb02474f506e43f4" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.474860 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xt6j5" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.953297 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7958b68d9f-qk798"] Sep 30 14:40:29 crc kubenswrapper[4799]: E0930 14:40:29.961964 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d30b6ec-a2d1-4f93-932b-e78856253cc9" containerName="barbican-db-sync" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.961997 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d30b6ec-a2d1-4f93-932b-e78856253cc9" containerName="barbican-db-sync" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.962262 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d30b6ec-a2d1-4f93-932b-e78856253cc9" containerName="barbican-db-sync" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.963803 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.967435 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.967737 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8vn4g" Sep 30 14:40:29 crc kubenswrapper[4799]: I0930 14:40:29.973313 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.015707 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5888b5f87b-kmzgb"] Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.017736 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.022501 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.069768 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7958b68d9f-qk798"] Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.114151 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5888b5f87b-kmzgb"] Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.142779 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-config-data-custom\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.142842 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e704a0e9-df6f-46c9-a9fd-8e5547031469-logs\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.142886 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-combined-ca-bundle\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.142908 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-config-data-custom\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.142940 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g6dd\" (UniqueName: \"kubernetes.io/projected/663d1769-2ce0-436c-986e-b07abcb30dca-kube-api-access-9g6dd\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.142970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/663d1769-2ce0-436c-986e-b07abcb30dca-logs\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.142984 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-combined-ca-bundle\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.143018 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-config-data\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.143064 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-config-data\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.143112 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95cjp\" (UniqueName: \"kubernetes.io/projected/e704a0e9-df6f-46c9-a9fd-8e5547031469-kube-api-access-95cjp\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.241439 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-sqctp"] Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.252167 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.253892 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95cjp\" (UniqueName: \"kubernetes.io/projected/e704a0e9-df6f-46c9-a9fd-8e5547031469-kube-api-access-95cjp\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.253987 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-config-data-custom\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.254044 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e704a0e9-df6f-46c9-a9fd-8e5547031469-logs\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.254099 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-combined-ca-bundle\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.254123 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-config-data-custom\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.254170 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g6dd\" (UniqueName: \"kubernetes.io/projected/663d1769-2ce0-436c-986e-b07abcb30dca-kube-api-access-9g6dd\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.254211 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/663d1769-2ce0-436c-986e-b07abcb30dca-logs\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.254236 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-combined-ca-bundle\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.254285 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-config-data\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.254348 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-config-data\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.260524 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/663d1769-2ce0-436c-986e-b07abcb30dca-logs\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.268485 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e704a0e9-df6f-46c9-a9fd-8e5547031469-logs\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.280694 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-combined-ca-bundle\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.285212 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-sqctp"] Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.339329 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-config-data\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.347005 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-config-data-custom\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.348096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-config-data-custom\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.398323 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e704a0e9-df6f-46c9-a9fd-8e5547031469-config-data\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.399124 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.399207 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.399241 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-config\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.399284 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.399360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8bsw\" (UniqueName: \"kubernetes.io/projected/2e206fe9-8c90-4d5d-8ecd-648160782ece-kube-api-access-c8bsw\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.399388 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.403472 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95cjp\" (UniqueName: \"kubernetes.io/projected/e704a0e9-df6f-46c9-a9fd-8e5547031469-kube-api-access-95cjp\") pod \"barbican-worker-7958b68d9f-qk798\" (UID: \"e704a0e9-df6f-46c9-a9fd-8e5547031469\") " pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.400841 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/663d1769-2ce0-436c-986e-b07abcb30dca-combined-ca-bundle\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.413665 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g6dd\" (UniqueName: \"kubernetes.io/projected/663d1769-2ce0-436c-986e-b07abcb30dca-kube-api-access-9g6dd\") pod \"barbican-keystone-listener-5888b5f87b-kmzgb\" (UID: \"663d1769-2ce0-436c-986e-b07abcb30dca\") " pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.453484 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7958b68d9f-qk798" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.489679 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.518861 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8bsw\" (UniqueName: \"kubernetes.io/projected/2e206fe9-8c90-4d5d-8ecd-648160782ece-kube-api-access-c8bsw\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.518942 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.519283 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.519374 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.519401 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-config\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.519465 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.519918 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6f7d8cd7cd-xsv2g"] Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.521777 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.521801 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.523260 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f7d8cd7cd-xsv2g"] Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.523515 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.524069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.524241 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-config\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.531020 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.539676 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.588777 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8bsw\" (UniqueName: \"kubernetes.io/projected/2e206fe9-8c90-4d5d-8ecd-648160782ece-kube-api-access-c8bsw\") pod \"dnsmasq-dns-75c8ddd69c-sqctp\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.733474 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data-custom\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.733950 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a558095-8ee7-4653-82de-9813105b4871-logs\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.734055 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6shd\" (UniqueName: \"kubernetes.io/projected/9a558095-8ee7-4653-82de-9813105b4871-kube-api-access-r6shd\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.734226 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-combined-ca-bundle\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.734327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.801837 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.837293 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-combined-ca-bundle\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.837392 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.837439 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data-custom\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.837597 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a558095-8ee7-4653-82de-9813105b4871-logs\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.837637 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6shd\" (UniqueName: \"kubernetes.io/projected/9a558095-8ee7-4653-82de-9813105b4871-kube-api-access-r6shd\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.840544 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a558095-8ee7-4653-82de-9813105b4871-logs\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.869053 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.871066 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6shd\" (UniqueName: \"kubernetes.io/projected/9a558095-8ee7-4653-82de-9813105b4871-kube-api-access-r6shd\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.872580 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data-custom\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:30 crc kubenswrapper[4799]: I0930 14:40:30.881980 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-combined-ca-bundle\") pod \"barbican-api-6f7d8cd7cd-xsv2g\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:31 crc kubenswrapper[4799]: I0930 14:40:31.156220 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:32 crc kubenswrapper[4799]: I0930 14:40:32.583168 4799 generic.go:334] "Generic (PLEG): container finished" podID="78ac655a-3fc7-46a4-93c2-69574e8b8f22" containerID="a86ae59bf3e2020b35f1021225e2b995c371d1ee708c374d3038fd248b6dcc86" exitCode=0 Sep 30 14:40:32 crc kubenswrapper[4799]: I0930 14:40:32.583440 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-mx9fz" event={"ID":"78ac655a-3fc7-46a4-93c2-69574e8b8f22","Type":"ContainerDied","Data":"a86ae59bf3e2020b35f1021225e2b995c371d1ee708c374d3038fd248b6dcc86"} Sep 30 14:40:33 crc kubenswrapper[4799]: W0930 14:40:33.361602 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e206fe9_8c90_4d5d_8ecd_648160782ece.slice/crio-5a8756e92255fd5d6e1193ecf92739c0d35dbb0c0428cd7d388b031911eb6ef1 WatchSource:0}: Error finding container 5a8756e92255fd5d6e1193ecf92739c0d35dbb0c0428cd7d388b031911eb6ef1: Status 404 returned error can't find the container with id 5a8756e92255fd5d6e1193ecf92739c0d35dbb0c0428cd7d388b031911eb6ef1 Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.364919 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-sqctp"] Sep 30 14:40:33 crc kubenswrapper[4799]: E0930 14:40:33.432359 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.496516 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f7d8cd7cd-xsv2g"] Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.669134 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" event={"ID":"2e206fe9-8c90-4d5d-8ecd-648160782ece","Type":"ContainerStarted","Data":"5a8756e92255fd5d6e1193ecf92739c0d35dbb0c0428cd7d388b031911eb6ef1"} Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.692335 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="ceilometer-notification-agent" containerID="cri-o://1dcb7f3d5d8af2b3a16c5b094fa265e2b8dccf34c2420b99e5d60c425be29877" gracePeriod=30 Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.692710 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc98dba3-94f3-4cc0-91b4-152a3fa006b7","Type":"ContainerStarted","Data":"1893bafe681069d24a62fd52542b13fb71918b5e00e91d505d076461b0e088d8"} Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.692764 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.693081 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="proxy-httpd" containerID="cri-o://1893bafe681069d24a62fd52542b13fb71918b5e00e91d505d076461b0e088d8" gracePeriod=30 Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.693138 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="sg-core" containerID="cri-o://7c8d2e81c5226bd8ee43d610fff5f34cde4c34313c054a0b2707edc9c2be2379" gracePeriod=30 Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.761559 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7958b68d9f-qk798"] Sep 30 14:40:33 crc kubenswrapper[4799]: I0930 14:40:33.786977 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5888b5f87b-kmzgb"] Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.243336 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-84f5689c94-tlpq6"] Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.269620 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-84f5689c94-tlpq6"] Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.269780 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.277816 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.279060 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.361415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94dzc\" (UniqueName: \"kubernetes.io/projected/b992c9d0-6ae3-4032-a4d5-f893c4813931-kube-api-access-94dzc\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.361530 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-internal-tls-certs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.361556 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-public-tls-certs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.361609 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b992c9d0-6ae3-4032-a4d5-f893c4813931-logs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.361634 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-combined-ca-bundle\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.361689 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-config-data\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.361804 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-config-data-custom\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.391500 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.409453 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.470639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94dzc\" (UniqueName: \"kubernetes.io/projected/b992c9d0-6ae3-4032-a4d5-f893c4813931-kube-api-access-94dzc\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.471128 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-internal-tls-certs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.471157 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-public-tls-certs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.471198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b992c9d0-6ae3-4032-a4d5-f893c4813931-logs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.471228 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-combined-ca-bundle\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.471267 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-config-data\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.471358 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-config-data-custom\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.472381 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b992c9d0-6ae3-4032-a4d5-f893c4813931-logs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.478802 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-config-data-custom\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.485668 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-combined-ca-bundle\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.492141 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-internal-tls-certs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.492280 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-public-tls-certs\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.501505 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94dzc\" (UniqueName: \"kubernetes.io/projected/b992c9d0-6ae3-4032-a4d5-f893c4813931-kube-api-access-94dzc\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.505950 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b992c9d0-6ae3-4032-a4d5-f893c4813931-config-data\") pod \"barbican-api-84f5689c94-tlpq6\" (UID: \"b992c9d0-6ae3-4032-a4d5-f893c4813931\") " pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.575686 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2txv\" (UniqueName: \"kubernetes.io/projected/78ac655a-3fc7-46a4-93c2-69574e8b8f22-kube-api-access-k2txv\") pod \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.575843 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-combined-ca-bundle\") pod \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.576015 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-config-data\") pod \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.576104 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-db-sync-config-data\") pod \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.576275 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ac655a-3fc7-46a4-93c2-69574e8b8f22-etc-machine-id\") pod \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.576336 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-scripts\") pod \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\" (UID: \"78ac655a-3fc7-46a4-93c2-69574e8b8f22\") " Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.585960 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78ac655a-3fc7-46a4-93c2-69574e8b8f22-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "78ac655a-3fc7-46a4-93c2-69574e8b8f22" (UID: "78ac655a-3fc7-46a4-93c2-69574e8b8f22"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.642405 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.657142 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-scripts" (OuterVolumeSpecName: "scripts") pod "78ac655a-3fc7-46a4-93c2-69574e8b8f22" (UID: "78ac655a-3fc7-46a4-93c2-69574e8b8f22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.831052 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78ac655a-3fc7-46a4-93c2-69574e8b8f22-kube-api-access-k2txv" (OuterVolumeSpecName: "kube-api-access-k2txv") pod "78ac655a-3fc7-46a4-93c2-69574e8b8f22" (UID: "78ac655a-3fc7-46a4-93c2-69574e8b8f22"). InnerVolumeSpecName "kube-api-access-k2txv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.842549 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "78ac655a-3fc7-46a4-93c2-69574e8b8f22" (UID: "78ac655a-3fc7-46a4-93c2-69574e8b8f22"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.853976 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78ac655a-3fc7-46a4-93c2-69574e8b8f22" (UID: "78ac655a-3fc7-46a4-93c2-69574e8b8f22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.883076 4799 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.883104 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78ac655a-3fc7-46a4-93c2-69574e8b8f22-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.883115 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.883132 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2txv\" (UniqueName: \"kubernetes.io/projected/78ac655a-3fc7-46a4-93c2-69574e8b8f22-kube-api-access-k2txv\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.883144 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.933139 4799 generic.go:334] "Generic (PLEG): container finished" podID="2e206fe9-8c90-4d5d-8ecd-648160782ece" containerID="574ee13417b257a11d1cd5644a5e0b0d6b13c3376680d9f7e25bce93f3bb8602" exitCode=0 Sep 30 14:40:34 crc kubenswrapper[4799]: I0930 14:40:34.934440 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" event={"ID":"2e206fe9-8c90-4d5d-8ecd-648160782ece","Type":"ContainerDied","Data":"574ee13417b257a11d1cd5644a5e0b0d6b13c3376680d9f7e25bce93f3bb8602"} Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.045776 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" event={"ID":"9a558095-8ee7-4653-82de-9813105b4871","Type":"ContainerStarted","Data":"b0f016ec7bb0a09705bc5e1b5a8dc013f9a87019812f4a15ab7770496ffd56d5"} Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.046141 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" event={"ID":"9a558095-8ee7-4653-82de-9813105b4871","Type":"ContainerStarted","Data":"f588430fa41ec469f8dab1a2912996bb0e3067b79af21822c9d6ed075bf58e73"} Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.046152 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" event={"ID":"9a558095-8ee7-4653-82de-9813105b4871","Type":"ContainerStarted","Data":"9f58d5ec4c89265b094c32cd90aa1d8ad9212d7f5e1b582b31a71b7bd04970b1"} Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.047911 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.052625 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.070285 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7958b68d9f-qk798" event={"ID":"e704a0e9-df6f-46c9-a9fd-8e5547031469","Type":"ContainerStarted","Data":"b937f43350a3db730841d5ff4bd02c5a7a69097e7d5570748c2d9b3c3dc7d0ca"} Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.097996 4799 generic.go:334] "Generic (PLEG): container finished" podID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerID="7c8d2e81c5226bd8ee43d610fff5f34cde4c34313c054a0b2707edc9c2be2379" exitCode=2 Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.098180 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc98dba3-94f3-4cc0-91b4-152a3fa006b7","Type":"ContainerDied","Data":"7c8d2e81c5226bd8ee43d610fff5f34cde4c34313c054a0b2707edc9c2be2379"} Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.106097 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" event={"ID":"663d1769-2ce0-436c-986e-b07abcb30dca","Type":"ContainerStarted","Data":"811a560132a1a78fd904f8b473e34b35e824d6ba1dc652e76bbb30c87a4dcfac"} Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.135500 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-mx9fz" event={"ID":"78ac655a-3fc7-46a4-93c2-69574e8b8f22","Type":"ContainerDied","Data":"3f86766104c25656916fcf4c1cfadbc141c5990594f7857b7ef2f2c800149e12"} Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.135566 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f86766104c25656916fcf4c1cfadbc141c5990594f7857b7ef2f2c800149e12" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.135687 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-mx9fz" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.155550 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:40:35 crc kubenswrapper[4799]: E0930 14:40:35.156393 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ac655a-3fc7-46a4-93c2-69574e8b8f22" containerName="cinder-db-sync" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.156472 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ac655a-3fc7-46a4-93c2-69574e8b8f22" containerName="cinder-db-sync" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.161136 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="78ac655a-3fc7-46a4-93c2-69574e8b8f22" containerName="cinder-db-sync" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.162796 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.215601 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.231017 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-config-data" (OuterVolumeSpecName: "config-data") pod "78ac655a-3fc7-46a4-93c2-69574e8b8f22" (UID: "78ac655a-3fc7-46a4-93c2-69574e8b8f22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.278765 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.291611 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podStartSLOduration=5.291583763 podStartE2EDuration="5.291583763s" podCreationTimestamp="2025-09-30 14:40:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:35.121752288 +0000 UTC m=+1257.205352715" watchObservedRunningTime="2025-09-30 14:40:35.291583763 +0000 UTC m=+1257.375184190" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.326759 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-sqctp"] Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.328369 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.328488 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.328531 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-scripts\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.328693 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.328718 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.328794 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sws5t\" (UniqueName: \"kubernetes.io/projected/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-kube-api-access-sws5t\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.328872 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78ac655a-3fc7-46a4-93c2-69574e8b8f22-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.413826 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jdtxl"] Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.416569 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.436443 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.436762 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.436996 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.436919 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sws5t\" (UniqueName: \"kubernetes.io/projected/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-kube-api-access-sws5t\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.437628 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.437867 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.437961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-scripts\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.454708 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jdtxl"] Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.465850 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.480803 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.481588 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.489245 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-scripts\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.491811 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sws5t\" (UniqueName: \"kubernetes.io/projected/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-kube-api-access-sws5t\") pod \"cinder-scheduler-0\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.545125 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-svc\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.545624 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljnxj\" (UniqueName: \"kubernetes.io/projected/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-kube-api-access-ljnxj\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.547780 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.547985 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.548187 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.548349 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-config\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.562612 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.578047 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.583757 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.591106 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.650951 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651035 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-svc\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651068 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd8z5\" (UniqueName: \"kubernetes.io/projected/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-kube-api-access-xd8z5\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651208 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data-custom\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651243 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651285 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljnxj\" (UniqueName: \"kubernetes.io/projected/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-kube-api-access-ljnxj\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651307 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651341 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651451 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651598 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-config\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651645 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-logs\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651679 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.651702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-scripts\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.652681 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-svc\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.658429 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-config\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.662270 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.662384 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.663679 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.668746 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.751715 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljnxj\" (UniqueName: \"kubernetes.io/projected/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-kube-api-access-ljnxj\") pod \"dnsmasq-dns-5784cf869f-jdtxl\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.754171 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-logs\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.754342 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.754416 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-scripts\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.754501 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.754640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd8z5\" (UniqueName: \"kubernetes.io/projected/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-kube-api-access-xd8z5\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.754794 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data-custom\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.754864 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.760145 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-logs\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.761806 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.778859 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.789152 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-84f5689c94-tlpq6"] Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.832351 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-scripts\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.832401 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data-custom\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.835820 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd8z5\" (UniqueName: \"kubernetes.io/projected/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-kube-api-access-xd8z5\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.895716 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.915239 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:40:35 crc kubenswrapper[4799]: I0930 14:40:35.920082 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:36 crc kubenswrapper[4799]: E0930 14:40:36.034756 4799 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Sep 30 14:40:36 crc kubenswrapper[4799]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/2e206fe9-8c90-4d5d-8ecd-648160782ece/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Sep 30 14:40:36 crc kubenswrapper[4799]: > podSandboxID="5a8756e92255fd5d6e1193ecf92739c0d35dbb0c0428cd7d388b031911eb6ef1" Sep 30 14:40:36 crc kubenswrapper[4799]: E0930 14:40:36.035322 4799 kuberuntime_manager.go:1274] "Unhandled Error" err=< Sep 30 14:40:36 crc kubenswrapper[4799]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n566h86hd4h5f9hc8h599h5h56bh75h554h597h5f4hb7h98h58fh66ch57ch668h5bfhd8h596h68dh54h8ch674h587h5bdhb9hc4h695h5b8hccq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c8bsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-75c8ddd69c-sqctp_openstack(2e206fe9-8c90-4d5d-8ecd-648160782ece): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/2e206fe9-8c90-4d5d-8ecd-648160782ece/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Sep 30 14:40:36 crc kubenswrapper[4799]: > logger="UnhandledError" Sep 30 14:40:36 crc kubenswrapper[4799]: E0930 14:40:36.036462 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/2e206fe9-8c90-4d5d-8ecd-648160782ece/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" podUID="2e206fe9-8c90-4d5d-8ecd-648160782ece" Sep 30 14:40:36 crc kubenswrapper[4799]: I0930 14:40:36.161865 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84f5689c94-tlpq6" event={"ID":"b992c9d0-6ae3-4032-a4d5-f893c4813931","Type":"ContainerStarted","Data":"55a5e6328a6926ed833495d85b5df011d9ee8f30af020c3b046d4d2bc45d1ed6"} Sep 30 14:40:36 crc kubenswrapper[4799]: I0930 14:40:36.468537 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:40:36 crc kubenswrapper[4799]: I0930 14:40:36.923634 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.022254 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.139314 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-swift-storage-0\") pod \"2e206fe9-8c90-4d5d-8ecd-648160782ece\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.139365 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-nb\") pod \"2e206fe9-8c90-4d5d-8ecd-648160782ece\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.139407 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-config\") pod \"2e206fe9-8c90-4d5d-8ecd-648160782ece\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.139568 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-sb\") pod \"2e206fe9-8c90-4d5d-8ecd-648160782ece\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.139641 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8bsw\" (UniqueName: \"kubernetes.io/projected/2e206fe9-8c90-4d5d-8ecd-648160782ece-kube-api-access-c8bsw\") pod \"2e206fe9-8c90-4d5d-8ecd-648160782ece\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.139676 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-svc\") pod \"2e206fe9-8c90-4d5d-8ecd-648160782ece\" (UID: \"2e206fe9-8c90-4d5d-8ecd-648160782ece\") " Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.201643 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e206fe9-8c90-4d5d-8ecd-648160782ece-kube-api-access-c8bsw" (OuterVolumeSpecName: "kube-api-access-c8bsw") pod "2e206fe9-8c90-4d5d-8ecd-648160782ece" (UID: "2e206fe9-8c90-4d5d-8ecd-648160782ece"). InnerVolumeSpecName "kube-api-access-c8bsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.225378 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81","Type":"ContainerStarted","Data":"08aa5767c6937a30960e65f9847c42b9c6d65919f93f3b742798f2ec851445ca"} Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.240468 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" event={"ID":"2e206fe9-8c90-4d5d-8ecd-648160782ece","Type":"ContainerDied","Data":"5a8756e92255fd5d6e1193ecf92739c0d35dbb0c0428cd7d388b031911eb6ef1"} Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.240537 4799 scope.go:117] "RemoveContainer" containerID="574ee13417b257a11d1cd5644a5e0b0d6b13c3376680d9f7e25bce93f3bb8602" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.240793 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-sqctp" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.250358 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84f5689c94-tlpq6" event={"ID":"b992c9d0-6ae3-4032-a4d5-f893c4813931","Type":"ContainerStarted","Data":"3a369925e11a86ef99673355885668ea0e61bde8f05a88ac916c600596bdb92b"} Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.255325 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8bsw\" (UniqueName: \"kubernetes.io/projected/2e206fe9-8c90-4d5d-8ecd-648160782ece-kube-api-access-c8bsw\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.259244 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bea5eda1-2610-4d17-a4d0-f8f414ca4e36","Type":"ContainerStarted","Data":"87140490e396d694bc85c545572ddce3155c84bda5f2bac4ae2af22c5bd627da"} Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.351842 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-config" (OuterVolumeSpecName: "config") pod "2e206fe9-8c90-4d5d-8ecd-648160782ece" (UID: "2e206fe9-8c90-4d5d-8ecd-648160782ece"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.368513 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jdtxl"] Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.394320 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2e206fe9-8c90-4d5d-8ecd-648160782ece" (UID: "2e206fe9-8c90-4d5d-8ecd-648160782ece"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.409846 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2e206fe9-8c90-4d5d-8ecd-648160782ece" (UID: "2e206fe9-8c90-4d5d-8ecd-648160782ece"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.435897 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2e206fe9-8c90-4d5d-8ecd-648160782ece" (UID: "2e206fe9-8c90-4d5d-8ecd-648160782ece"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.439258 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.439472 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:37 crc kubenswrapper[4799]: W0930 14:40:37.445004 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7720d30e_281a_4f4f_be9c_0b8f1bc475f6.slice/crio-6a3d416cda5549819559da5d14b45d38267f8d91e2ec42e028ddd5525e819079 WatchSource:0}: Error finding container 6a3d416cda5549819559da5d14b45d38267f8d91e2ec42e028ddd5525e819079: Status 404 returned error can't find the container with id 6a3d416cda5549819559da5d14b45d38267f8d91e2ec42e028ddd5525e819079 Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.542383 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.568835 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.579473 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.633489 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2e206fe9-8c90-4d5d-8ecd-648160782ece" (UID: "2e206fe9-8c90-4d5d-8ecd-648160782ece"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.684460 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e206fe9-8c90-4d5d-8ecd-648160782ece-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.682938 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.685777 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.687507 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"0db7b3053e7c7468a8690c8b47ee3b620779c70dce53efef8b358e92bdd7e983"} pod="openstack/horizon-7f7c888686-pwv2q" containerMessage="Container horizon failed startup probe, will be restarted" Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.693313 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" containerID="cri-o://0db7b3053e7c7468a8690c8b47ee3b620779c70dce53efef8b358e92bdd7e983" gracePeriod=30 Sep 30 14:40:37 crc kubenswrapper[4799]: I0930 14:40:37.992724 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-sqctp"] Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.018192 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-sqctp"] Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.295539 4799 generic.go:334] "Generic (PLEG): container finished" podID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerID="1dcb7f3d5d8af2b3a16c5b094fa265e2b8dccf34c2420b99e5d60c425be29877" exitCode=0 Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.295621 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc98dba3-94f3-4cc0-91b4-152a3fa006b7","Type":"ContainerDied","Data":"1dcb7f3d5d8af2b3a16c5b094fa265e2b8dccf34c2420b99e5d60c425be29877"} Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.299949 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84f5689c94-tlpq6" event={"ID":"b992c9d0-6ae3-4032-a4d5-f893c4813931","Type":"ContainerStarted","Data":"2c34916e8f619f15444d42a4118ba7bacce87e82a1b11fb193e7a67010db3fd1"} Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.300732 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.300892 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.314463 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" event={"ID":"7720d30e-281a-4f4f-be9c-0b8f1bc475f6","Type":"ContainerStarted","Data":"6a3d416cda5549819559da5d14b45d38267f8d91e2ec42e028ddd5525e819079"} Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.329667 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-84f5689c94-tlpq6" podStartSLOduration=4.329610306 podStartE2EDuration="4.329610306s" podCreationTimestamp="2025-09-30 14:40:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:38.320396557 +0000 UTC m=+1260.403997004" watchObservedRunningTime="2025-09-30 14:40:38.329610306 +0000 UTC m=+1260.413210743" Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.562171 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e206fe9-8c90-4d5d-8ecd-648160782ece" path="/var/lib/kubelet/pods/2e206fe9-8c90-4d5d-8ecd-648160782ece/volumes" Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.749489 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-67c5b7d44f-24zhg" Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.798634 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.866394 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-664b88d46b-sllrg"] Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.874115 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-664b88d46b-sllrg" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerName="neutron-api" containerID="cri-o://6609957b639734329472ddec0bca06f450dfeaaa8fb6132e950d36c48aea6e4a" gracePeriod=30 Sep 30 14:40:38 crc kubenswrapper[4799]: I0930 14:40:38.874766 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-664b88d46b-sllrg" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerName="neutron-httpd" containerID="cri-o://d4173870f24110850c8b8a5559df233eebbb65a87542709008409ecb3e009c79" gracePeriod=30 Sep 30 14:40:39 crc kubenswrapper[4799]: I0930 14:40:39.340906 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bea5eda1-2610-4d17-a4d0-f8f414ca4e36","Type":"ContainerStarted","Data":"79e8acef280cb68e39e41873c10e53df851a710dd0869fec41715054d849d2e2"} Sep 30 14:40:40 crc kubenswrapper[4799]: I0930 14:40:40.361828 4799 generic.go:334] "Generic (PLEG): container finished" podID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" containerID="5cb5e6829b3afc23b8e4bc1f2df067901aa9c895fc1897c6133fb7845aa14fe7" exitCode=0 Sep 30 14:40:40 crc kubenswrapper[4799]: I0930 14:40:40.362788 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" event={"ID":"7720d30e-281a-4f4f-be9c-0b8f1bc475f6","Type":"ContainerDied","Data":"5cb5e6829b3afc23b8e4bc1f2df067901aa9c895fc1897c6133fb7845aa14fe7"} Sep 30 14:40:40 crc kubenswrapper[4799]: I0930 14:40:40.373843 4799 generic.go:334] "Generic (PLEG): container finished" podID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerID="d4173870f24110850c8b8a5559df233eebbb65a87542709008409ecb3e009c79" exitCode=0 Sep 30 14:40:40 crc kubenswrapper[4799]: I0930 14:40:40.373912 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-664b88d46b-sllrg" event={"ID":"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150","Type":"ContainerDied","Data":"d4173870f24110850c8b8a5559df233eebbb65a87542709008409ecb3e009c79"} Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.571013 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" event={"ID":"663d1769-2ce0-436c-986e-b07abcb30dca","Type":"ContainerStarted","Data":"ddc814b9b3148b5b289d03ad6b1b770ff377550560395f59e402267e90d84cab"} Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.571562 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" event={"ID":"663d1769-2ce0-436c-986e-b07abcb30dca","Type":"ContainerStarted","Data":"48797b05c4d624187f874bc2c48ec578001b4cfa1cdfca4406cb3f3daf9dd6d7"} Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.582370 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81","Type":"ContainerStarted","Data":"ca5455e66babe15974405972e99ac9c895bb6244b20eaf4b26e6eff3bfeef1ae"} Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.596374 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" event={"ID":"7720d30e-281a-4f4f-be9c-0b8f1bc475f6","Type":"ContainerStarted","Data":"0dbe8d62cbbbfc49cb340f4e23a3440773b4a36fb9035d93f4596e3bb41f36fd"} Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.600850 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5888b5f87b-kmzgb" podStartSLOduration=6.373702644 podStartE2EDuration="13.60080124s" podCreationTimestamp="2025-09-30 14:40:29 +0000 UTC" firstStartedPulling="2025-09-30 14:40:33.849252489 +0000 UTC m=+1255.932852926" lastFinishedPulling="2025-09-30 14:40:41.076351085 +0000 UTC m=+1263.159951522" observedRunningTime="2025-09-30 14:40:42.58779738 +0000 UTC m=+1264.671397827" watchObservedRunningTime="2025-09-30 14:40:42.60080124 +0000 UTC m=+1264.684401677" Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.603263 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.637694 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7958b68d9f-qk798" event={"ID":"e704a0e9-df6f-46c9-a9fd-8e5547031469","Type":"ContainerStarted","Data":"4a1e652eec086685343426c8cf9221bc926891ac0535710d25622ade2a096bb4"} Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.637752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7958b68d9f-qk798" event={"ID":"e704a0e9-df6f-46c9-a9fd-8e5547031469","Type":"ContainerStarted","Data":"b1b4b8085c8a0280e67a3a80835a63017827437c0a2dd5892f0ff66eae02d13b"} Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.640959 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" podStartSLOduration=7.640931603 podStartE2EDuration="7.640931603s" podCreationTimestamp="2025-09-30 14:40:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:42.639161341 +0000 UTC m=+1264.722761778" watchObservedRunningTime="2025-09-30 14:40:42.640931603 +0000 UTC m=+1264.724532030" Sep 30 14:40:42 crc kubenswrapper[4799]: I0930 14:40:42.706892 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7958b68d9f-qk798" podStartSLOduration=6.470396831 podStartE2EDuration="13.7068375s" podCreationTimestamp="2025-09-30 14:40:29 +0000 UTC" firstStartedPulling="2025-09-30 14:40:33.849795675 +0000 UTC m=+1255.933396102" lastFinishedPulling="2025-09-30 14:40:41.086236344 +0000 UTC m=+1263.169836771" observedRunningTime="2025-09-30 14:40:42.675293858 +0000 UTC m=+1264.758894285" watchObservedRunningTime="2025-09-30 14:40:42.7068375 +0000 UTC m=+1264.790437927" Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.686103 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81","Type":"ContainerStarted","Data":"ee7dbc11a2167a6b8c44212feb14e459bab750c619900accafde8ac16b3f97cc"} Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.721996 4799 generic.go:334] "Generic (PLEG): container finished" podID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerID="6609957b639734329472ddec0bca06f450dfeaaa8fb6132e950d36c48aea6e4a" exitCode=0 Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.722168 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-664b88d46b-sllrg" event={"ID":"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150","Type":"ContainerDied","Data":"6609957b639734329472ddec0bca06f450dfeaaa8fb6132e950d36c48aea6e4a"} Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.726282 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.321470312 podStartE2EDuration="8.726246691s" podCreationTimestamp="2025-09-30 14:40:35 +0000 UTC" firstStartedPulling="2025-09-30 14:40:36.571177141 +0000 UTC m=+1258.654777568" lastFinishedPulling="2025-09-30 14:40:40.97595352 +0000 UTC m=+1263.059553947" observedRunningTime="2025-09-30 14:40:43.722131441 +0000 UTC m=+1265.805731908" watchObservedRunningTime="2025-09-30 14:40:43.726246691 +0000 UTC m=+1265.809847128" Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.746865 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api-log" containerID="cri-o://79e8acef280cb68e39e41873c10e53df851a710dd0869fec41715054d849d2e2" gracePeriod=30 Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.747310 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bea5eda1-2610-4d17-a4d0-f8f414ca4e36","Type":"ContainerStarted","Data":"65552e8a3afc408f710ddffabc74117115242eee1d3516faf95d4d73e80aafcf"} Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.751932 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api" containerID="cri-o://65552e8a3afc408f710ddffabc74117115242eee1d3516faf95d4d73e80aafcf" gracePeriod=30 Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.752560 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 30 14:40:43 crc kubenswrapper[4799]: I0930 14:40:43.792963 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=8.79290206 podStartE2EDuration="8.79290206s" podCreationTimestamp="2025-09-30 14:40:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:40:43.790133459 +0000 UTC m=+1265.873733886" watchObservedRunningTime="2025-09-30 14:40:43.79290206 +0000 UTC m=+1265.876502497" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.165672 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.328519 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-combined-ca-bundle\") pod \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.328905 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-ovndb-tls-certs\") pod \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.328999 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5j57\" (UniqueName: \"kubernetes.io/projected/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-kube-api-access-z5j57\") pod \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.329146 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-httpd-config\") pod \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.329883 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-config\") pod \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\" (UID: \"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150\") " Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.373805 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" (UID: "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.393485 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-kube-api-access-z5j57" (OuterVolumeSpecName: "kube-api-access-z5j57") pod "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" (UID: "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150"). InnerVolumeSpecName "kube-api-access-z5j57". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.436645 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5j57\" (UniqueName: \"kubernetes.io/projected/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-kube-api-access-z5j57\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.436954 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-httpd-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.482867 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-config" (OuterVolumeSpecName: "config") pod "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" (UID: "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.540088 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.577988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" (UID: "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.653524 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.686217 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" (UID: "0f4d4cd7-aa4d-4f61-be9d-848c9c72b150"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.758394 4799 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.774765 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-664b88d46b-sllrg" event={"ID":"0f4d4cd7-aa4d-4f61-be9d-848c9c72b150","Type":"ContainerDied","Data":"f6f84ddeb12c8e433a6cd607dcb07ac70d9dd3622d2231a3fe7019d9bd1e0f25"} Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.775225 4799 scope.go:117] "RemoveContainer" containerID="d4173870f24110850c8b8a5559df233eebbb65a87542709008409ecb3e009c79" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.775175 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-664b88d46b-sllrg" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.787337 4799 generic.go:334] "Generic (PLEG): container finished" podID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerID="79e8acef280cb68e39e41873c10e53df851a710dd0869fec41715054d849d2e2" exitCode=143 Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.788853 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bea5eda1-2610-4d17-a4d0-f8f414ca4e36","Type":"ContainerDied","Data":"79e8acef280cb68e39e41873c10e53df851a710dd0869fec41715054d849d2e2"} Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.820386 4799 scope.go:117] "RemoveContainer" containerID="6609957b639734329472ddec0bca06f450dfeaaa8fb6132e950d36c48aea6e4a" Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.850538 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-664b88d46b-sllrg"] Sep 30 14:40:44 crc kubenswrapper[4799]: I0930 14:40:44.866622 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-664b88d46b-sllrg"] Sep 30 14:40:45 crc kubenswrapper[4799]: I0930 14:40:45.241995 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:45 crc kubenswrapper[4799]: I0930 14:40:45.242683 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:45 crc kubenswrapper[4799]: I0930 14:40:45.563995 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 30 14:40:46 crc kubenswrapper[4799]: I0930 14:40:46.239958 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:46 crc kubenswrapper[4799]: I0930 14:40:46.239991 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:46 crc kubenswrapper[4799]: I0930 14:40:46.516555 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" path="/var/lib/kubelet/pods/0f4d4cd7-aa4d-4f61-be9d-848c9c72b150/volumes" Sep 30 14:40:46 crc kubenswrapper[4799]: I0930 14:40:46.541751 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 30 14:40:47 crc kubenswrapper[4799]: I0930 14:40:47.529443 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:40:47 crc kubenswrapper[4799]: I0930 14:40:47.529557 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:40:47 crc kubenswrapper[4799]: I0930 14:40:47.530627 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"6218e563ed5620135720c08ff067c9d79266769a1728d09cab51ff7af21c5dfe"} pod="openstack/horizon-7f99689854-glm7t" containerMessage="Container horizon failed startup probe, will be restarted" Sep 30 14:40:47 crc kubenswrapper[4799]: I0930 14:40:47.531015 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" containerID="cri-o://6218e563ed5620135720c08ff067c9d79266769a1728d09cab51ff7af21c5dfe" gracePeriod=30 Sep 30 14:40:48 crc kubenswrapper[4799]: I0930 14:40:48.656941 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-84f5689c94-tlpq6" podUID="b992c9d0-6ae3-4032-a4d5-f893c4813931" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:49 crc kubenswrapper[4799]: I0930 14:40:49.264312 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:49 crc kubenswrapper[4799]: I0930 14:40:49.264389 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-9fd6b988d-fckfn" Sep 30 14:40:49 crc kubenswrapper[4799]: I0930 14:40:49.668969 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-84f5689c94-tlpq6" podUID="b992c9d0-6ae3-4032-a4d5-f893c4813931" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:49 crc kubenswrapper[4799]: I0930 14:40:49.674854 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-84f5689c94-tlpq6" podUID="b992c9d0-6ae3-4032-a4d5-f893c4813931" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:50 crc kubenswrapper[4799]: I0930 14:40:50.334042 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:50 crc kubenswrapper[4799]: I0930 14:40:50.335059 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:50 crc kubenswrapper[4799]: I0930 14:40:50.567886 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.163:8080/\": dial tcp 10.217.0.163:8080: connect: connection refused" Sep 30 14:40:50 crc kubenswrapper[4799]: I0930 14:40:50.925822 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:40:51 crc kubenswrapper[4799]: I0930 14:40:51.023941 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-bzzsz"] Sep 30 14:40:51 crc kubenswrapper[4799]: I0930 14:40:51.024241 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" podUID="afa0ef8b-835f-4d37-b610-3261288185db" containerName="dnsmasq-dns" containerID="cri-o://e53d95b617c76885c77c6041547acd2f4243ccce407e260e09accbb601b17122" gracePeriod=10 Sep 30 14:40:51 crc kubenswrapper[4799]: I0930 14:40:51.323332 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:51 crc kubenswrapper[4799]: I0930 14:40:51.324468 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:51 crc kubenswrapper[4799]: I0930 14:40:51.343977 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:51 crc kubenswrapper[4799]: I0930 14:40:51.349125 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:51 crc kubenswrapper[4799]: I0930 14:40:51.901111 4799 generic.go:334] "Generic (PLEG): container finished" podID="afa0ef8b-835f-4d37-b610-3261288185db" containerID="e53d95b617c76885c77c6041547acd2f4243ccce407e260e09accbb601b17122" exitCode=0 Sep 30 14:40:51 crc kubenswrapper[4799]: I0930 14:40:51.902585 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" event={"ID":"afa0ef8b-835f-4d37-b610-3261288185db","Type":"ContainerDied","Data":"e53d95b617c76885c77c6041547acd2f4243ccce407e260e09accbb601b17122"} Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.037750 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.155457 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-config\") pod \"afa0ef8b-835f-4d37-b610-3261288185db\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.155535 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-sb\") pod \"afa0ef8b-835f-4d37-b610-3261288185db\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.155592 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-swift-storage-0\") pod \"afa0ef8b-835f-4d37-b610-3261288185db\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.155689 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-nb\") pod \"afa0ef8b-835f-4d37-b610-3261288185db\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.155821 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnc8c\" (UniqueName: \"kubernetes.io/projected/afa0ef8b-835f-4d37-b610-3261288185db-kube-api-access-qnc8c\") pod \"afa0ef8b-835f-4d37-b610-3261288185db\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.155864 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-svc\") pod \"afa0ef8b-835f-4d37-b610-3261288185db\" (UID: \"afa0ef8b-835f-4d37-b610-3261288185db\") " Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.169663 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa0ef8b-835f-4d37-b610-3261288185db-kube-api-access-qnc8c" (OuterVolumeSpecName: "kube-api-access-qnc8c") pod "afa0ef8b-835f-4d37-b610-3261288185db" (UID: "afa0ef8b-835f-4d37-b610-3261288185db"). InnerVolumeSpecName "kube-api-access-qnc8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.261412 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnc8c\" (UniqueName: \"kubernetes.io/projected/afa0ef8b-835f-4d37-b610-3261288185db-kube-api-access-qnc8c\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.306756 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "afa0ef8b-835f-4d37-b610-3261288185db" (UID: "afa0ef8b-835f-4d37-b610-3261288185db"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.315847 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "afa0ef8b-835f-4d37-b610-3261288185db" (UID: "afa0ef8b-835f-4d37-b610-3261288185db"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.337362 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "afa0ef8b-835f-4d37-b610-3261288185db" (UID: "afa0ef8b-835f-4d37-b610-3261288185db"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.340490 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "afa0ef8b-835f-4d37-b610-3261288185db" (UID: "afa0ef8b-835f-4d37-b610-3261288185db"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.361584 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-config" (OuterVolumeSpecName: "config") pod "afa0ef8b-835f-4d37-b610-3261288185db" (UID: "afa0ef8b-835f-4d37-b610-3261288185db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.363160 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.363197 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.363217 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.363233 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.363245 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa0ef8b-835f-4d37-b610-3261288185db-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.921430 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" event={"ID":"afa0ef8b-835f-4d37-b610-3261288185db","Type":"ContainerDied","Data":"a77a3c40f8cd9d1f517ff1c987583ac19ac814ba64f60640b6ddb0499023b961"} Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.921523 4799 scope.go:117] "RemoveContainer" containerID="e53d95b617c76885c77c6041547acd2f4243ccce407e260e09accbb601b17122" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.921822 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-bzzsz" Sep 30 14:40:52 crc kubenswrapper[4799]: I0930 14:40:52.957988 4799 scope.go:117] "RemoveContainer" containerID="46abd85e21d46e3e4bf3f246d44a5c1f83e78d0de2ae1f7e49b01ad61bb65083" Sep 30 14:40:53 crc kubenswrapper[4799]: I0930 14:40:53.004907 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-bzzsz"] Sep 30 14:40:53 crc kubenswrapper[4799]: I0930 14:40:53.006860 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-84f5689c94-tlpq6" podUID="b992c9d0-6ae3-4032-a4d5-f893c4813931" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:53 crc kubenswrapper[4799]: I0930 14:40:53.036465 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-bzzsz"] Sep 30 14:40:53 crc kubenswrapper[4799]: I0930 14:40:53.663030 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-84f5689c94-tlpq6" podUID="b992c9d0-6ae3-4032-a4d5-f893c4813931" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.400519 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.526387 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa0ef8b-835f-4d37-b610-3261288185db" path="/var/lib/kubelet/pods/afa0ef8b-835f-4d37-b610-3261288185db/volumes" Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.680930 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-84f5689c94-tlpq6" podUID="b992c9d0-6ae3-4032-a4d5-f893c4813931" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.694346 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-84f5689c94-tlpq6" Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.775948 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f7d8cd7cd-xsv2g"] Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.776288 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" containerID="cri-o://f588430fa41ec469f8dab1a2912996bb0e3067b79af21822c9d6ed075bf58e73" gracePeriod=30 Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.776866 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" containerID="cri-o://b0f016ec7bb0a09705bc5e1b5a8dc013f9a87019812f4a15ab7770496ffd56d5" gracePeriod=30 Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.955924 4799 generic.go:334] "Generic (PLEG): container finished" podID="9a558095-8ee7-4653-82de-9813105b4871" containerID="f588430fa41ec469f8dab1a2912996bb0e3067b79af21822c9d6ed075bf58e73" exitCode=143 Sep 30 14:40:54 crc kubenswrapper[4799]: I0930 14:40:54.956015 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" event={"ID":"9a558095-8ee7-4653-82de-9813105b4871","Type":"ContainerDied","Data":"f588430fa41ec469f8dab1a2912996bb0e3067b79af21822c9d6ed075bf58e73"} Sep 30 14:40:55 crc kubenswrapper[4799]: I0930 14:40:55.437538 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-577d747d48-ldwh6" Sep 30 14:40:55 crc kubenswrapper[4799]: I0930 14:40:55.958019 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.165:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:40:55 crc kubenswrapper[4799]: I0930 14:40:55.958201 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 30 14:40:56 crc kubenswrapper[4799]: I0930 14:40:56.049455 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:40:56 crc kubenswrapper[4799]: I0930 14:40:56.978188 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="cinder-scheduler" containerID="cri-o://ca5455e66babe15974405972e99ac9c895bb6244b20eaf4b26e6eff3bfeef1ae" gracePeriod=30 Sep 30 14:40:56 crc kubenswrapper[4799]: I0930 14:40:56.978282 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="probe" containerID="cri-o://ee7dbc11a2167a6b8c44212feb14e459bab750c619900accafde8ac16b3f97cc" gracePeriod=30 Sep 30 14:40:57 crc kubenswrapper[4799]: I0930 14:40:57.990395 4799 generic.go:334] "Generic (PLEG): container finished" podID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerID="ee7dbc11a2167a6b8c44212feb14e459bab750c619900accafde8ac16b3f97cc" exitCode=0 Sep 30 14:40:57 crc kubenswrapper[4799]: I0930 14:40:57.990508 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81","Type":"ContainerDied","Data":"ee7dbc11a2167a6b8c44212feb14e459bab750c619900accafde8ac16b3f97cc"} Sep 30 14:40:58 crc kubenswrapper[4799]: I0930 14:40:58.306635 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:36070->10.217.0.161:9311: read: connection reset by peer" Sep 30 14:40:58 crc kubenswrapper[4799]: I0930 14:40:58.307604 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:36068->10.217.0.161:9311: read: connection reset by peer" Sep 30 14:40:58 crc kubenswrapper[4799]: I0930 14:40:58.983283 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.026050 4799 generic.go:334] "Generic (PLEG): container finished" podID="9a558095-8ee7-4653-82de-9813105b4871" containerID="b0f016ec7bb0a09705bc5e1b5a8dc013f9a87019812f4a15ab7770496ffd56d5" exitCode=0 Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.026112 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" event={"ID":"9a558095-8ee7-4653-82de-9813105b4871","Type":"ContainerDied","Data":"b0f016ec7bb0a09705bc5e1b5a8dc013f9a87019812f4a15ab7770496ffd56d5"} Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.205304 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.353755 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data\") pod \"9a558095-8ee7-4653-82de-9813105b4871\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.353850 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a558095-8ee7-4653-82de-9813105b4871-logs\") pod \"9a558095-8ee7-4653-82de-9813105b4871\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.353896 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data-custom\") pod \"9a558095-8ee7-4653-82de-9813105b4871\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.353930 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6shd\" (UniqueName: \"kubernetes.io/projected/9a558095-8ee7-4653-82de-9813105b4871-kube-api-access-r6shd\") pod \"9a558095-8ee7-4653-82de-9813105b4871\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.353988 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-combined-ca-bundle\") pod \"9a558095-8ee7-4653-82de-9813105b4871\" (UID: \"9a558095-8ee7-4653-82de-9813105b4871\") " Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.354828 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a558095-8ee7-4653-82de-9813105b4871-logs" (OuterVolumeSpecName: "logs") pod "9a558095-8ee7-4653-82de-9813105b4871" (UID: "9a558095-8ee7-4653-82de-9813105b4871"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.430273 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9a558095-8ee7-4653-82de-9813105b4871" (UID: "9a558095-8ee7-4653-82de-9813105b4871"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.430298 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a558095-8ee7-4653-82de-9813105b4871-kube-api-access-r6shd" (OuterVolumeSpecName: "kube-api-access-r6shd") pod "9a558095-8ee7-4653-82de-9813105b4871" (UID: "9a558095-8ee7-4653-82de-9813105b4871"). InnerVolumeSpecName "kube-api-access-r6shd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.458419 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a558095-8ee7-4653-82de-9813105b4871-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.458467 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.458481 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6shd\" (UniqueName: \"kubernetes.io/projected/9a558095-8ee7-4653-82de-9813105b4871-kube-api-access-r6shd\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.466059 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a558095-8ee7-4653-82de-9813105b4871" (UID: "9a558095-8ee7-4653-82de-9813105b4871"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.530360 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data" (OuterVolumeSpecName: "config-data") pod "9a558095-8ee7-4653-82de-9813105b4871" (UID: "9a558095-8ee7-4653-82de-9813105b4871"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.563197 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.563260 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a558095-8ee7-4653-82de-9813105b4871-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.648519 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:40:59 crc kubenswrapper[4799]: I0930 14:40:59.648597 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.045930 4799 generic.go:334] "Generic (PLEG): container finished" podID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerID="ca5455e66babe15974405972e99ac9c895bb6244b20eaf4b26e6eff3bfeef1ae" exitCode=0 Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.045969 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81","Type":"ContainerDied","Data":"ca5455e66babe15974405972e99ac9c895bb6244b20eaf4b26e6eff3bfeef1ae"} Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.048910 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" event={"ID":"9a558095-8ee7-4653-82de-9813105b4871","Type":"ContainerDied","Data":"9f58d5ec4c89265b094c32cd90aa1d8ad9212d7f5e1b582b31a71b7bd04970b1"} Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.048956 4799 scope.go:117] "RemoveContainer" containerID="b0f016ec7bb0a09705bc5e1b5a8dc013f9a87019812f4a15ab7770496ffd56d5" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.049145 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f7d8cd7cd-xsv2g" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.128377 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f7d8cd7cd-xsv2g"] Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.139480 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6f7d8cd7cd-xsv2g"] Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.157216 4799 scope.go:117] "RemoveContainer" containerID="f588430fa41ec469f8dab1a2912996bb0e3067b79af21822c9d6ed075bf58e73" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.353779 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.354305 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354333 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.354350 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerName="neutron-api" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354359 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerName="neutron-api" Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.354380 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e206fe9-8c90-4d5d-8ecd-648160782ece" containerName="init" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354389 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e206fe9-8c90-4d5d-8ecd-648160782ece" containerName="init" Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.354409 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354417 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.354438 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerName="neutron-httpd" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354445 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerName="neutron-httpd" Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.354464 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa0ef8b-835f-4d37-b610-3261288185db" containerName="dnsmasq-dns" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354471 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa0ef8b-835f-4d37-b610-3261288185db" containerName="dnsmasq-dns" Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.354484 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa0ef8b-835f-4d37-b610-3261288185db" containerName="init" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354492 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa0ef8b-835f-4d37-b610-3261288185db" containerName="init" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354740 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354801 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerName="neutron-httpd" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354827 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a558095-8ee7-4653-82de-9813105b4871" containerName="barbican-api-log" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354886 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa0ef8b-835f-4d37-b610-3261288185db" containerName="dnsmasq-dns" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354908 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e206fe9-8c90-4d5d-8ecd-648160782ece" containerName="init" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.354980 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f4d4cd7-aa4d-4f61-be9d-848c9c72b150" containerName="neutron-api" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.356075 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.359431 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-6jm9s" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.363553 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.370799 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.384860 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.589408 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.610134 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a558095-8ee7-4653-82de-9813105b4871" path="/var/lib/kubelet/pods/9a558095-8ee7-4653-82de-9813105b4871/volumes" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.613165 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.613369 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config-secret\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.613584 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txvws\" (UniqueName: \"kubernetes.io/projected/d5d1f932-51f9-4247-ae52-d369b30efa48-kube-api-access-txvws\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.613840 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.695149 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.704110 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-txvws openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="d5d1f932-51f9-4247-ae52-d369b30efa48" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716064 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-combined-ca-bundle\") pod \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716135 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-scripts\") pod \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716224 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data-custom\") pod \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716276 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data\") pod \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716358 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-etc-machine-id\") pod \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sws5t\" (UniqueName: \"kubernetes.io/projected/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-kube-api-access-sws5t\") pod \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\" (UID: \"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81\") " Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716797 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txvws\" (UniqueName: \"kubernetes.io/projected/d5d1f932-51f9-4247-ae52-d369b30efa48-kube-api-access-txvws\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716887 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716917 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" (UID: "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.716961 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.717031 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config-secret\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.717104 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.718534 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.749546 4799 projected.go:194] Error preparing data for projected volume kube-api-access-txvws for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.750411 4799 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d5d1f932-51f9-4247-ae52-d369b30efa48-kube-api-access-txvws podName:d5d1f932-51f9-4247-ae52-d369b30efa48 nodeName:}" failed. No retries permitted until 2025-09-30 14:41:01.250378704 +0000 UTC m=+1283.333979131 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-txvws" (UniqueName: "kubernetes.io/projected/d5d1f932-51f9-4247-ae52-d369b30efa48-kube-api-access-txvws") pod "openstackclient" (UID: "d5d1f932-51f9-4247-ae52-d369b30efa48") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.754136 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.768045 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-kube-api-access-sws5t" (OuterVolumeSpecName: "kube-api-access-sws5t") pod "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" (UID: "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81"). InnerVolumeSpecName "kube-api-access-sws5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.768161 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-scripts" (OuterVolumeSpecName: "scripts") pod "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" (UID: "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.768248 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" (UID: "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.770164 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config-secret\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.799129 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.804852 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.805406 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="probe" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.805428 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="probe" Sep 30 14:41:00 crc kubenswrapper[4799]: E0930 14:41:00.805477 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="cinder-scheduler" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.805484 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="cinder-scheduler" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.805734 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="probe" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.805759 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" containerName="cinder-scheduler" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.806528 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.815766 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.818018 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/93dd541f-bdad-44fb-9450-1baa7acabb40-openstack-config\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.818079 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/93dd541f-bdad-44fb-9450-1baa7acabb40-openstack-config-secret\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.818107 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q7km\" (UniqueName: \"kubernetes.io/projected/93dd541f-bdad-44fb-9450-1baa7acabb40-kube-api-access-2q7km\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.818171 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93dd541f-bdad-44fb-9450-1baa7acabb40-combined-ca-bundle\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.818296 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sws5t\" (UniqueName: \"kubernetes.io/projected/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-kube-api-access-sws5t\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.818308 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.818319 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.920311 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/93dd541f-bdad-44fb-9450-1baa7acabb40-openstack-config-secret\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.920385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q7km\" (UniqueName: \"kubernetes.io/projected/93dd541f-bdad-44fb-9450-1baa7acabb40-kube-api-access-2q7km\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.920506 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93dd541f-bdad-44fb-9450-1baa7acabb40-combined-ca-bundle\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.920737 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/93dd541f-bdad-44fb-9450-1baa7acabb40-openstack-config\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.922304 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/93dd541f-bdad-44fb-9450-1baa7acabb40-openstack-config\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.923786 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" (UID: "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.928710 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93dd541f-bdad-44fb-9450-1baa7acabb40-combined-ca-bundle\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.931958 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/93dd541f-bdad-44fb-9450-1baa7acabb40-openstack-config-secret\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.950599 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q7km\" (UniqueName: \"kubernetes.io/projected/93dd541f-bdad-44fb-9450-1baa7acabb40-kube-api-access-2q7km\") pod \"openstackclient\" (UID: \"93dd541f-bdad-44fb-9450-1baa7acabb40\") " pod="openstack/openstackclient" Sep 30 14:41:00 crc kubenswrapper[4799]: I0930 14:41:00.973266 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data" (OuterVolumeSpecName: "config-data") pod "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" (UID: "223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.023448 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.023825 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.080451 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81","Type":"ContainerDied","Data":"08aa5767c6937a30960e65f9847c42b9c6d65919f93f3b742798f2ec851445ca"} Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.080604 4799 scope.go:117] "RemoveContainer" containerID="ee7dbc11a2167a6b8c44212feb14e459bab750c619900accafde8ac16b3f97cc" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.080830 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.114333 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.121237 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d5d1f932-51f9-4247-ae52-d369b30efa48" podUID="93dd541f-bdad-44fb-9450-1baa7acabb40" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.143730 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.195872 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.204765 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.216332 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.229013 4799 scope.go:117] "RemoveContainer" containerID="ca5455e66babe15974405972e99ac9c895bb6244b20eaf4b26e6eff3bfeef1ae" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.232036 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.234535 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.237290 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.240855 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.334544 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config-secret\") pod \"d5d1f932-51f9-4247-ae52-d369b30efa48\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.334713 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-combined-ca-bundle\") pod \"d5d1f932-51f9-4247-ae52-d369b30efa48\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.334832 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config\") pod \"d5d1f932-51f9-4247-ae52-d369b30efa48\" (UID: \"d5d1f932-51f9-4247-ae52-d369b30efa48\") " Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.335418 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txvws\" (UniqueName: \"kubernetes.io/projected/d5d1f932-51f9-4247-ae52-d369b30efa48-kube-api-access-txvws\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.335946 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d5d1f932-51f9-4247-ae52-d369b30efa48" (UID: "d5d1f932-51f9-4247-ae52-d369b30efa48"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.340629 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d5d1f932-51f9-4247-ae52-d369b30efa48" (UID: "d5d1f932-51f9-4247-ae52-d369b30efa48"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.345870 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5d1f932-51f9-4247-ae52-d369b30efa48" (UID: "d5d1f932-51f9-4247-ae52-d369b30efa48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.437194 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.437263 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-config-data\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.437382 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.438278 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n277b\" (UniqueName: \"kubernetes.io/projected/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-kube-api-access-n277b\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.438359 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-scripts\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.438391 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.438474 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.438493 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d1f932-51f9-4247-ae52-d369b30efa48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.438505 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5d1f932-51f9-4247-ae52-d369b30efa48-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.541203 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n277b\" (UniqueName: \"kubernetes.io/projected/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-kube-api-access-n277b\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.541565 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-scripts\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.541625 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.541794 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.541830 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-config-data\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.541929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.542201 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.554483 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.563097 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-scripts\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.564913 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-config-data\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.567204 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.579466 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n277b\" (UniqueName: \"kubernetes.io/projected/5c94a3cf-a75d-469a-9c13-ca9dd2ff978b-kube-api-access-n277b\") pod \"cinder-scheduler-0\" (UID: \"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b\") " pod="openstack/cinder-scheduler-0" Sep 30 14:41:01 crc kubenswrapper[4799]: I0930 14:41:01.860925 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Sep 30 14:41:02 crc kubenswrapper[4799]: I0930 14:41:02.096636 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Sep 30 14:41:02 crc kubenswrapper[4799]: W0930 14:41:02.148431 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93dd541f_bdad_44fb_9450_1baa7acabb40.slice/crio-c52107f9a4f282ae1c4bfbfa81b6b3b323ca87fa7cfd07220d65101a0f94284c WatchSource:0}: Error finding container c52107f9a4f282ae1c4bfbfa81b6b3b323ca87fa7cfd07220d65101a0f94284c: Status 404 returned error can't find the container with id c52107f9a4f282ae1c4bfbfa81b6b3b323ca87fa7cfd07220d65101a0f94284c Sep 30 14:41:02 crc kubenswrapper[4799]: I0930 14:41:02.156267 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Sep 30 14:41:02 crc kubenswrapper[4799]: I0930 14:41:02.168947 4799 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="d5d1f932-51f9-4247-ae52-d369b30efa48" podUID="93dd541f-bdad-44fb-9450-1baa7acabb40" Sep 30 14:41:02 crc kubenswrapper[4799]: E0930 14:41:02.319432 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5d1f932_51f9_4247_ae52_d369b30efa48.slice\": RecentStats: unable to find data in memory cache]" Sep 30 14:41:02 crc kubenswrapper[4799]: I0930 14:41:02.523892 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81" path="/var/lib/kubelet/pods/223dfbc4-19f1-4cf9-b2fb-b5d9e740ac81/volumes" Sep 30 14:41:02 crc kubenswrapper[4799]: I0930 14:41:02.524804 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5d1f932-51f9-4247-ae52-d369b30efa48" path="/var/lib/kubelet/pods/d5d1f932-51f9-4247-ae52-d369b30efa48/volumes" Sep 30 14:41:02 crc kubenswrapper[4799]: I0930 14:41:02.578573 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Sep 30 14:41:03 crc kubenswrapper[4799]: I0930 14:41:03.172191 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b","Type":"ContainerStarted","Data":"0fd5fb2ebcb222e273a4c2ef792623846977233f79327103b7b5f360943629af"} Sep 30 14:41:03 crc kubenswrapper[4799]: I0930 14:41:03.184690 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"93dd541f-bdad-44fb-9450-1baa7acabb40","Type":"ContainerStarted","Data":"c52107f9a4f282ae1c4bfbfa81b6b3b323ca87fa7cfd07220d65101a0f94284c"} Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.203800 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b","Type":"ContainerStarted","Data":"451ed6dffc684cd7b8cea64a8244309aac6a73ca1c5a4b538de8b692c53af5fa"} Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.218026 4799 generic.go:334] "Generic (PLEG): container finished" podID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerID="1893bafe681069d24a62fd52542b13fb71918b5e00e91d505d076461b0e088d8" exitCode=137 Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.218114 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc98dba3-94f3-4cc0-91b4-152a3fa006b7","Type":"ContainerDied","Data":"1893bafe681069d24a62fd52542b13fb71918b5e00e91d505d076461b0e088d8"} Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.575318 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.735603 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-log-httpd\") pod \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.736129 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8nrj\" (UniqueName: \"kubernetes.io/projected/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-kube-api-access-f8nrj\") pod \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.736238 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bc98dba3-94f3-4cc0-91b4-152a3fa006b7" (UID: "bc98dba3-94f3-4cc0-91b4-152a3fa006b7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.736306 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-run-httpd\") pod \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.736363 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-config-data\") pod \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.736429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-scripts\") pod \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.736506 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-sg-core-conf-yaml\") pod \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.736533 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-combined-ca-bundle\") pod \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\" (UID: \"bc98dba3-94f3-4cc0-91b4-152a3fa006b7\") " Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.736799 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bc98dba3-94f3-4cc0-91b4-152a3fa006b7" (UID: "bc98dba3-94f3-4cc0-91b4-152a3fa006b7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.740837 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.740876 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.746864 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-scripts" (OuterVolumeSpecName: "scripts") pod "bc98dba3-94f3-4cc0-91b4-152a3fa006b7" (UID: "bc98dba3-94f3-4cc0-91b4-152a3fa006b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.750031 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-kube-api-access-f8nrj" (OuterVolumeSpecName: "kube-api-access-f8nrj") pod "bc98dba3-94f3-4cc0-91b4-152a3fa006b7" (UID: "bc98dba3-94f3-4cc0-91b4-152a3fa006b7"). InnerVolumeSpecName "kube-api-access-f8nrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.799828 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bc98dba3-94f3-4cc0-91b4-152a3fa006b7" (UID: "bc98dba3-94f3-4cc0-91b4-152a3fa006b7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.854877 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc98dba3-94f3-4cc0-91b4-152a3fa006b7" (UID: "bc98dba3-94f3-4cc0-91b4-152a3fa006b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.855027 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-config-data" (OuterVolumeSpecName: "config-data") pod "bc98dba3-94f3-4cc0-91b4-152a3fa006b7" (UID: "bc98dba3-94f3-4cc0-91b4-152a3fa006b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.856815 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.856865 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8nrj\" (UniqueName: \"kubernetes.io/projected/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-kube-api-access-f8nrj\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.856880 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.856892 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:04 crc kubenswrapper[4799]: I0930 14:41:04.856904 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc98dba3-94f3-4cc0-91b4-152a3fa006b7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.231560 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc98dba3-94f3-4cc0-91b4-152a3fa006b7","Type":"ContainerDied","Data":"732391f9f2193e29ff38a4a58268c34c98aad094fc22cc7167811aaf3fdfc5f4"} Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.231589 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.231690 4799 scope.go:117] "RemoveContainer" containerID="1893bafe681069d24a62fd52542b13fb71918b5e00e91d505d076461b0e088d8" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.237803 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5c94a3cf-a75d-469a-9c13-ca9dd2ff978b","Type":"ContainerStarted","Data":"fdfbafcaeef0eff0106a0f33217a169aadd0af5eeecb4b1542f7c9e2ffdf0e4a"} Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.268985 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.268959849 podStartE2EDuration="4.268959849s" podCreationTimestamp="2025-09-30 14:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:41:05.257876595 +0000 UTC m=+1287.341477032" watchObservedRunningTime="2025-09-30 14:41:05.268959849 +0000 UTC m=+1287.352560266" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.291288 4799 scope.go:117] "RemoveContainer" containerID="7c8d2e81c5226bd8ee43d610fff5f34cde4c34313c054a0b2707edc9c2be2379" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.315922 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.320761 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.349988 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:05 crc kubenswrapper[4799]: E0930 14:41:05.350486 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="ceilometer-notification-agent" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.350663 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="ceilometer-notification-agent" Sep 30 14:41:05 crc kubenswrapper[4799]: E0930 14:41:05.350696 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="sg-core" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.350705 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="sg-core" Sep 30 14:41:05 crc kubenswrapper[4799]: E0930 14:41:05.350720 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="proxy-httpd" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.350726 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="proxy-httpd" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.350912 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="sg-core" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.350939 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="ceilometer-notification-agent" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.350951 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" containerName="proxy-httpd" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.352778 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.361791 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.362288 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.381328 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-log-httpd\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.381397 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-run-httpd\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.381421 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-scripts\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.381450 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.381487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76ldv\" (UniqueName: \"kubernetes.io/projected/fefc41cf-7c0f-43f9-a41b-01412bd697a6-kube-api-access-76ldv\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.381517 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.381606 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-config-data\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.395752 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.414956 4799 scope.go:117] "RemoveContainer" containerID="1dcb7f3d5d8af2b3a16c5b094fa265e2b8dccf34c2420b99e5d60c425be29877" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.482977 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.483385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-config-data\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.483524 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-log-httpd\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.483569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-run-httpd\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.483588 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-scripts\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.483625 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.483687 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76ldv\" (UniqueName: \"kubernetes.io/projected/fefc41cf-7c0f-43f9-a41b-01412bd697a6-kube-api-access-76ldv\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.484517 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-log-httpd\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.488463 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-run-httpd\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.492731 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-config-data\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.494937 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-scripts\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.505231 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.505790 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.506262 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76ldv\" (UniqueName: \"kubernetes.io/projected/fefc41cf-7c0f-43f9-a41b-01412bd697a6-kube-api-access-76ldv\") pod \"ceilometer-0\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " pod="openstack/ceilometer-0" Sep 30 14:41:05 crc kubenswrapper[4799]: I0930 14:41:05.680402 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:06 crc kubenswrapper[4799]: I0930 14:41:06.421177 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:06 crc kubenswrapper[4799]: W0930 14:41:06.448971 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfefc41cf_7c0f_43f9_a41b_01412bd697a6.slice/crio-bff9cd2de5d0f2e07155548661ae9f7eb9c35d66947a6676b0803f6bcd807ea3 WatchSource:0}: Error finding container bff9cd2de5d0f2e07155548661ae9f7eb9c35d66947a6676b0803f6bcd807ea3: Status 404 returned error can't find the container with id bff9cd2de5d0f2e07155548661ae9f7eb9c35d66947a6676b0803f6bcd807ea3 Sep 30 14:41:06 crc kubenswrapper[4799]: I0930 14:41:06.521514 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc98dba3-94f3-4cc0-91b4-152a3fa006b7" path="/var/lib/kubelet/pods/bc98dba3-94f3-4cc0-91b4-152a3fa006b7/volumes" Sep 30 14:41:06 crc kubenswrapper[4799]: I0930 14:41:06.862064 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.266690 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerStarted","Data":"bff9cd2de5d0f2e07155548661ae9f7eb9c35d66947a6676b0803f6bcd807ea3"} Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.518191 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5cfc9f87d7-nj7t8"] Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.531716 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.544089 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.544840 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.545022 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.576119 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5cfc9f87d7-nj7t8"] Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.635526 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-public-tls-certs\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.635669 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-internal-tls-certs\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.635725 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-combined-ca-bundle\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.635754 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-config-data\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.635778 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f32c852d-173a-4417-a268-38b7a480937e-run-httpd\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.635828 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfbqz\" (UniqueName: \"kubernetes.io/projected/f32c852d-173a-4417-a268-38b7a480937e-kube-api-access-gfbqz\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.635858 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f32c852d-173a-4417-a268-38b7a480937e-log-httpd\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.635925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f32c852d-173a-4417-a268-38b7a480937e-etc-swift\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.738303 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-internal-tls-certs\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.738411 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-combined-ca-bundle\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.738465 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-config-data\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.738494 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f32c852d-173a-4417-a268-38b7a480937e-run-httpd\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.738554 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfbqz\" (UniqueName: \"kubernetes.io/projected/f32c852d-173a-4417-a268-38b7a480937e-kube-api-access-gfbqz\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.738603 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f32c852d-173a-4417-a268-38b7a480937e-log-httpd\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.738819 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f32c852d-173a-4417-a268-38b7a480937e-etc-swift\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.738930 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-public-tls-certs\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.739774 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f32c852d-173a-4417-a268-38b7a480937e-run-httpd\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.740826 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f32c852d-173a-4417-a268-38b7a480937e-log-httpd\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.745395 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-combined-ca-bundle\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.746139 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-config-data\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.757269 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f32c852d-173a-4417-a268-38b7a480937e-etc-swift\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.763487 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-internal-tls-certs\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.770517 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfbqz\" (UniqueName: \"kubernetes.io/projected/f32c852d-173a-4417-a268-38b7a480937e-kube-api-access-gfbqz\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.773848 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f32c852d-173a-4417-a268-38b7a480937e-public-tls-certs\") pod \"swift-proxy-5cfc9f87d7-nj7t8\" (UID: \"f32c852d-173a-4417-a268-38b7a480937e\") " pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:07 crc kubenswrapper[4799]: I0930 14:41:07.878623 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:08 crc kubenswrapper[4799]: I0930 14:41:08.293374 4799 generic.go:334] "Generic (PLEG): container finished" podID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerID="0db7b3053e7c7468a8690c8b47ee3b620779c70dce53efef8b358e92bdd7e983" exitCode=137 Sep 30 14:41:08 crc kubenswrapper[4799]: I0930 14:41:08.293750 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerDied","Data":"0db7b3053e7c7468a8690c8b47ee3b620779c70dce53efef8b358e92bdd7e983"} Sep 30 14:41:08 crc kubenswrapper[4799]: W0930 14:41:08.772207 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf32c852d_173a_4417_a268_38b7a480937e.slice/crio-acf324b0832b0ae50221e4503299d0dd545b6ef502686dd7f508464ebcc7cbda WatchSource:0}: Error finding container acf324b0832b0ae50221e4503299d0dd545b6ef502686dd7f508464ebcc7cbda: Status 404 returned error can't find the container with id acf324b0832b0ae50221e4503299d0dd545b6ef502686dd7f508464ebcc7cbda Sep 30 14:41:08 crc kubenswrapper[4799]: I0930 14:41:08.785914 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5cfc9f87d7-nj7t8"] Sep 30 14:41:09 crc kubenswrapper[4799]: I0930 14:41:09.343736 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" event={"ID":"f32c852d-173a-4417-a268-38b7a480937e","Type":"ContainerStarted","Data":"acf324b0832b0ae50221e4503299d0dd545b6ef502686dd7f508464ebcc7cbda"} Sep 30 14:41:09 crc kubenswrapper[4799]: I0930 14:41:09.359161 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerStarted","Data":"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd"} Sep 30 14:41:09 crc kubenswrapper[4799]: I0930 14:41:09.376057 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerStarted","Data":"2aa2de6ee9c44af7d31b1105ea4177bb893946376162ccd5d747ef4d0d8305e3"} Sep 30 14:41:09 crc kubenswrapper[4799]: I0930 14:41:09.877027 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:10 crc kubenswrapper[4799]: I0930 14:41:10.391957 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" event={"ID":"f32c852d-173a-4417-a268-38b7a480937e","Type":"ContainerStarted","Data":"392998c82958f582b5c4f607682a6dfbb6d3971b314e1c7200583042093a0c48"} Sep 30 14:41:10 crc kubenswrapper[4799]: I0930 14:41:10.392055 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" event={"ID":"f32c852d-173a-4417-a268-38b7a480937e","Type":"ContainerStarted","Data":"85fdaa47b5db672af39a4686bbc8dc4cf9978334825705bcef19ff300bd43dda"} Sep 30 14:41:10 crc kubenswrapper[4799]: I0930 14:41:10.392499 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:10 crc kubenswrapper[4799]: I0930 14:41:10.392600 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:10 crc kubenswrapper[4799]: I0930 14:41:10.409079 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerStarted","Data":"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49"} Sep 30 14:41:10 crc kubenswrapper[4799]: I0930 14:41:10.409129 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerStarted","Data":"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86"} Sep 30 14:41:10 crc kubenswrapper[4799]: I0930 14:41:10.439600 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" podStartSLOduration=3.439571336 podStartE2EDuration="3.439571336s" podCreationTimestamp="2025-09-30 14:41:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:41:10.421058914 +0000 UTC m=+1292.504659371" watchObservedRunningTime="2025-09-30 14:41:10.439571336 +0000 UTC m=+1292.523171763" Sep 30 14:41:12 crc kubenswrapper[4799]: I0930 14:41:12.367061 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Sep 30 14:41:14 crc kubenswrapper[4799]: I0930 14:41:14.488331 4799 generic.go:334] "Generic (PLEG): container finished" podID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerID="65552e8a3afc408f710ddffabc74117115242eee1d3516faf95d4d73e80aafcf" exitCode=137 Sep 30 14:41:14 crc kubenswrapper[4799]: I0930 14:41:14.488428 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bea5eda1-2610-4d17-a4d0-f8f414ca4e36","Type":"ContainerDied","Data":"65552e8a3afc408f710ddffabc74117115242eee1d3516faf95d4d73e80aafcf"} Sep 30 14:41:15 crc kubenswrapper[4799]: I0930 14:41:15.915997 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.165:8776/healthcheck\": dial tcp 10.217.0.165:8776: connect: connection refused" Sep 30 14:41:17 crc kubenswrapper[4799]: I0930 14:41:17.681910 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:41:17 crc kubenswrapper[4799]: I0930 14:41:17.683860 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:41:17 crc kubenswrapper[4799]: I0930 14:41:17.892352 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:17 crc kubenswrapper[4799]: I0930 14:41:17.892431 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" Sep 30 14:41:18 crc kubenswrapper[4799]: I0930 14:41:18.541315 4799 generic.go:334] "Generic (PLEG): container finished" podID="d15890bb-d088-4126-8382-b5d8e423b608" containerID="6218e563ed5620135720c08ff067c9d79266769a1728d09cab51ff7af21c5dfe" exitCode=137 Sep 30 14:41:18 crc kubenswrapper[4799]: I0930 14:41:18.541338 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerDied","Data":"6218e563ed5620135720c08ff067c9d79266769a1728d09cab51ff7af21c5dfe"} Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.765338 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.798566 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd8z5\" (UniqueName: \"kubernetes.io/projected/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-kube-api-access-xd8z5\") pod \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.798638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data\") pod \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.798688 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-scripts\") pod \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.798717 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-etc-machine-id\") pod \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.798879 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bea5eda1-2610-4d17-a4d0-f8f414ca4e36" (UID: "bea5eda1-2610-4d17-a4d0-f8f414ca4e36"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.799494 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data-custom\") pod \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.799558 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-logs\") pod \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.799700 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-combined-ca-bundle\") pod \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\" (UID: \"bea5eda1-2610-4d17-a4d0-f8f414ca4e36\") " Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.800197 4799 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-etc-machine-id\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.801254 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-logs" (OuterVolumeSpecName: "logs") pod "bea5eda1-2610-4d17-a4d0-f8f414ca4e36" (UID: "bea5eda1-2610-4d17-a4d0-f8f414ca4e36"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.808390 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-kube-api-access-xd8z5" (OuterVolumeSpecName: "kube-api-access-xd8z5") pod "bea5eda1-2610-4d17-a4d0-f8f414ca4e36" (UID: "bea5eda1-2610-4d17-a4d0-f8f414ca4e36"). InnerVolumeSpecName "kube-api-access-xd8z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.813885 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bea5eda1-2610-4d17-a4d0-f8f414ca4e36" (UID: "bea5eda1-2610-4d17-a4d0-f8f414ca4e36"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.818696 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-scripts" (OuterVolumeSpecName: "scripts") pod "bea5eda1-2610-4d17-a4d0-f8f414ca4e36" (UID: "bea5eda1-2610-4d17-a4d0-f8f414ca4e36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.857268 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bea5eda1-2610-4d17-a4d0-f8f414ca4e36" (UID: "bea5eda1-2610-4d17-a4d0-f8f414ca4e36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.893713 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data" (OuterVolumeSpecName: "config-data") pod "bea5eda1-2610-4d17-a4d0-f8f414ca4e36" (UID: "bea5eda1-2610-4d17-a4d0-f8f414ca4e36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.904355 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.904420 4799 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data-custom\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.904439 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.904449 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.904460 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd8z5\" (UniqueName: \"kubernetes.io/projected/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-kube-api-access-xd8z5\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:20 crc kubenswrapper[4799]: I0930 14:41:20.904473 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bea5eda1-2610-4d17-a4d0-f8f414ca4e36-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.631806 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerStarted","Data":"2e60bf33011afb729f10179f6e079df29dfb26c28596d7ab55738fe40269d6c7"} Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.642913 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bea5eda1-2610-4d17-a4d0-f8f414ca4e36","Type":"ContainerDied","Data":"87140490e396d694bc85c545572ddce3155c84bda5f2bac4ae2af22c5bd627da"} Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.642991 4799 scope.go:117] "RemoveContainer" containerID="65552e8a3afc408f710ddffabc74117115242eee1d3516faf95d4d73e80aafcf" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.643202 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.646308 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"93dd541f-bdad-44fb-9450-1baa7acabb40","Type":"ContainerStarted","Data":"26805323bf203ae58426ea81d72ce9a4d90ced796c4ba37b2c14e7c0ba41d0fe"} Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.651932 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerStarted","Data":"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb"} Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.671746 4799 scope.go:117] "RemoveContainer" containerID="79e8acef280cb68e39e41873c10e53df851a710dd0869fec41715054d849d2e2" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.688071 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.714078 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.738413 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:41:21 crc kubenswrapper[4799]: E0930 14:41:21.739052 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api-log" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.739075 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api-log" Sep 30 14:41:21 crc kubenswrapper[4799]: E0930 14:41:21.739092 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.739099 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.742149 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.742188 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" containerName="cinder-api-log" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.743574 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.747252 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.750259 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.750544 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.750689 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.834215 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.834782 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppzfb\" (UniqueName: \"kubernetes.io/projected/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-kube-api-access-ppzfb\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.834867 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-config-data-custom\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.835052 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-scripts\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.835219 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.835297 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-logs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.835453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-config-data\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.835683 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.835822 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.937914 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.937999 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-config-data\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938024 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-logs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938028 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938125 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938166 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938264 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppzfb\" (UniqueName: \"kubernetes.io/projected/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-kube-api-access-ppzfb\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938283 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-config-data-custom\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938341 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-scripts\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.938497 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-logs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.944289 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-config-data\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.950012 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-scripts\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.950782 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.953379 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.954294 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.962141 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppzfb\" (UniqueName: \"kubernetes.io/projected/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-kube-api-access-ppzfb\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:21 crc kubenswrapper[4799]: I0930 14:41:21.963044 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da52c15f-1f9f-4485-99bc-1ea6b29c94c8-config-data-custom\") pod \"cinder-api-0\" (UID: \"da52c15f-1f9f-4485-99bc-1ea6b29c94c8\") " pod="openstack/cinder-api-0" Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.080878 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.523425 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bea5eda1-2610-4d17-a4d0-f8f414ca4e36" path="/var/lib/kubelet/pods/bea5eda1-2610-4d17-a4d0-f8f414ca4e36/volumes" Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.650755 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Sep 30 14:41:22 crc kubenswrapper[4799]: W0930 14:41:22.672322 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda52c15f_1f9f_4485_99bc_1ea6b29c94c8.slice/crio-25a6027e3098aa11093b6ed2d48a0f39b67e00de4731d9ac0df24302c746c5ca WatchSource:0}: Error finding container 25a6027e3098aa11093b6ed2d48a0f39b67e00de4731d9ac0df24302c746c5ca: Status 404 returned error can't find the container with id 25a6027e3098aa11093b6ed2d48a0f39b67e00de4731d9ac0df24302c746c5ca Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.677721 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="ceilometer-central-agent" containerID="cri-o://c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd" gracePeriod=30 Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.677861 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="proxy-httpd" containerID="cri-o://25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb" gracePeriod=30 Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.677922 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="sg-core" containerID="cri-o://d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49" gracePeriod=30 Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.677976 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="ceilometer-notification-agent" containerID="cri-o://e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86" gracePeriod=30 Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.710893 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.724838819 podStartE2EDuration="22.710866143s" podCreationTimestamp="2025-09-30 14:41:00 +0000 UTC" firstStartedPulling="2025-09-30 14:41:02.157005465 +0000 UTC m=+1284.240605892" lastFinishedPulling="2025-09-30 14:41:21.143032789 +0000 UTC m=+1303.226633216" observedRunningTime="2025-09-30 14:41:22.704297201 +0000 UTC m=+1304.787897628" watchObservedRunningTime="2025-09-30 14:41:22.710866143 +0000 UTC m=+1304.794466570" Sep 30 14:41:22 crc kubenswrapper[4799]: I0930 14:41:22.742471 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.055157332 podStartE2EDuration="17.742442517s" podCreationTimestamp="2025-09-30 14:41:05 +0000 UTC" firstStartedPulling="2025-09-30 14:41:06.453975152 +0000 UTC m=+1288.537575579" lastFinishedPulling="2025-09-30 14:41:21.141260337 +0000 UTC m=+1303.224860764" observedRunningTime="2025-09-30 14:41:22.739567002 +0000 UTC m=+1304.823167449" watchObservedRunningTime="2025-09-30 14:41:22.742442517 +0000 UTC m=+1304.826042944" Sep 30 14:41:23 crc kubenswrapper[4799]: E0930 14:41:23.136741 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfefc41cf_7c0f_43f9_a41b_01412bd697a6.slice/crio-conmon-d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfefc41cf_7c0f_43f9_a41b_01412bd697a6.slice/crio-conmon-25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfefc41cf_7c0f_43f9_a41b_01412bd697a6.slice/crio-25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb.scope\": RecentStats: unable to find data in memory cache]" Sep 30 14:41:23 crc kubenswrapper[4799]: I0930 14:41:23.694418 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da52c15f-1f9f-4485-99bc-1ea6b29c94c8","Type":"ContainerStarted","Data":"25a6027e3098aa11093b6ed2d48a0f39b67e00de4731d9ac0df24302c746c5ca"} Sep 30 14:41:23 crc kubenswrapper[4799]: I0930 14:41:23.698284 4799 generic.go:334] "Generic (PLEG): container finished" podID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerID="25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb" exitCode=0 Sep 30 14:41:23 crc kubenswrapper[4799]: I0930 14:41:23.698321 4799 generic.go:334] "Generic (PLEG): container finished" podID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerID="d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49" exitCode=2 Sep 30 14:41:23 crc kubenswrapper[4799]: I0930 14:41:23.698342 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerDied","Data":"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb"} Sep 30 14:41:23 crc kubenswrapper[4799]: I0930 14:41:23.698369 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerDied","Data":"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49"} Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.504247 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.566417 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-run-httpd\") pod \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.569959 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fefc41cf-7c0f-43f9-a41b-01412bd697a6" (UID: "fefc41cf-7c0f-43f9-a41b-01412bd697a6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.572405 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.673766 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76ldv\" (UniqueName: \"kubernetes.io/projected/fefc41cf-7c0f-43f9-a41b-01412bd697a6-kube-api-access-76ldv\") pod \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.673857 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-log-httpd\") pod \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.673890 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-combined-ca-bundle\") pod \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.673932 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-sg-core-conf-yaml\") pod \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.674163 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-config-data\") pod \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.674196 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-scripts\") pod \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\" (UID: \"fefc41cf-7c0f-43f9-a41b-01412bd697a6\") " Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.677288 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fefc41cf-7c0f-43f9-a41b-01412bd697a6" (UID: "fefc41cf-7c0f-43f9-a41b-01412bd697a6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.680781 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-scripts" (OuterVolumeSpecName: "scripts") pod "fefc41cf-7c0f-43f9-a41b-01412bd697a6" (UID: "fefc41cf-7c0f-43f9-a41b-01412bd697a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.686361 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fefc41cf-7c0f-43f9-a41b-01412bd697a6-kube-api-access-76ldv" (OuterVolumeSpecName: "kube-api-access-76ldv") pod "fefc41cf-7c0f-43f9-a41b-01412bd697a6" (UID: "fefc41cf-7c0f-43f9-a41b-01412bd697a6"). InnerVolumeSpecName "kube-api-access-76ldv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.723846 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da52c15f-1f9f-4485-99bc-1ea6b29c94c8","Type":"ContainerStarted","Data":"7e29ebcc3cb5617d61b66a4e035ae47885d4ccbdac4787f19dfecbb2a5a39e65"} Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.730579 4799 generic.go:334] "Generic (PLEG): container finished" podID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerID="e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86" exitCode=0 Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.730615 4799 generic.go:334] "Generic (PLEG): container finished" podID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerID="c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd" exitCode=0 Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.730638 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerDied","Data":"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86"} Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.730682 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerDied","Data":"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd"} Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.730692 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fefc41cf-7c0f-43f9-a41b-01412bd697a6","Type":"ContainerDied","Data":"bff9cd2de5d0f2e07155548661ae9f7eb9c35d66947a6676b0803f6bcd807ea3"} Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.730710 4799 scope.go:117] "RemoveContainer" containerID="25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.730873 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.733716 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fefc41cf-7c0f-43f9-a41b-01412bd697a6" (UID: "fefc41cf-7c0f-43f9-a41b-01412bd697a6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.778751 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76ldv\" (UniqueName: \"kubernetes.io/projected/fefc41cf-7c0f-43f9-a41b-01412bd697a6-kube-api-access-76ldv\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.778803 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fefc41cf-7c0f-43f9-a41b-01412bd697a6-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.778818 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.778830 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.811506 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fefc41cf-7c0f-43f9-a41b-01412bd697a6" (UID: "fefc41cf-7c0f-43f9-a41b-01412bd697a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.870800 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-config-data" (OuterVolumeSpecName: "config-data") pod "fefc41cf-7c0f-43f9-a41b-01412bd697a6" (UID: "fefc41cf-7c0f-43f9-a41b-01412bd697a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.880251 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.880290 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fefc41cf-7c0f-43f9-a41b-01412bd697a6-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.960036 4799 scope.go:117] "RemoveContainer" containerID="d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49" Sep 30 14:41:24 crc kubenswrapper[4799]: I0930 14:41:24.984292 4799 scope.go:117] "RemoveContainer" containerID="e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.009698 4799 scope.go:117] "RemoveContainer" containerID="c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.053867 4799 scope.go:117] "RemoveContainer" containerID="25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb" Sep 30 14:41:25 crc kubenswrapper[4799]: E0930 14:41:25.056975 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb\": container with ID starting with 25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb not found: ID does not exist" containerID="25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.057027 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb"} err="failed to get container status \"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb\": rpc error: code = NotFound desc = could not find container \"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb\": container with ID starting with 25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb not found: ID does not exist" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.057076 4799 scope.go:117] "RemoveContainer" containerID="d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49" Sep 30 14:41:25 crc kubenswrapper[4799]: E0930 14:41:25.060967 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49\": container with ID starting with d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49 not found: ID does not exist" containerID="d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.061028 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49"} err="failed to get container status \"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49\": rpc error: code = NotFound desc = could not find container \"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49\": container with ID starting with d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49 not found: ID does not exist" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.061083 4799 scope.go:117] "RemoveContainer" containerID="e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86" Sep 30 14:41:25 crc kubenswrapper[4799]: E0930 14:41:25.061445 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86\": container with ID starting with e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86 not found: ID does not exist" containerID="e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.061470 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86"} err="failed to get container status \"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86\": rpc error: code = NotFound desc = could not find container \"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86\": container with ID starting with e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86 not found: ID does not exist" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.061496 4799 scope.go:117] "RemoveContainer" containerID="c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd" Sep 30 14:41:25 crc kubenswrapper[4799]: E0930 14:41:25.063029 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd\": container with ID starting with c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd not found: ID does not exist" containerID="c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.063062 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd"} err="failed to get container status \"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd\": rpc error: code = NotFound desc = could not find container \"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd\": container with ID starting with c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd not found: ID does not exist" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.063081 4799 scope.go:117] "RemoveContainer" containerID="25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.067188 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb"} err="failed to get container status \"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb\": rpc error: code = NotFound desc = could not find container \"25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb\": container with ID starting with 25437f8bda3dc92bc7c6123011211b43789ba27d06299af9f387e67db7d4e8cb not found: ID does not exist" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.067250 4799 scope.go:117] "RemoveContainer" containerID="d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.067740 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49"} err="failed to get container status \"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49\": rpc error: code = NotFound desc = could not find container \"d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49\": container with ID starting with d5c2972682805efbf4fa274f9e49201280b06204829071528f6fbdfabd6b5d49 not found: ID does not exist" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.067758 4799 scope.go:117] "RemoveContainer" containerID="e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.071817 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86"} err="failed to get container status \"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86\": rpc error: code = NotFound desc = could not find container \"e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86\": container with ID starting with e38d495171dc66b2d607afe8386e28e61dcfc69014e7b6982623201f960acb86 not found: ID does not exist" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.071857 4799 scope.go:117] "RemoveContainer" containerID="c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.073685 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd"} err="failed to get container status \"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd\": rpc error: code = NotFound desc = could not find container \"c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd\": container with ID starting with c94f58ab64d602750a7ec7e62871628a6c2821fa0a01db00c844fcb0b170b5fd not found: ID does not exist" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.077663 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.088133 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.104270 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:25 crc kubenswrapper[4799]: E0930 14:41:25.104807 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="ceilometer-notification-agent" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.104835 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="ceilometer-notification-agent" Sep 30 14:41:25 crc kubenswrapper[4799]: E0930 14:41:25.104876 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="ceilometer-central-agent" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.104884 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="ceilometer-central-agent" Sep 30 14:41:25 crc kubenswrapper[4799]: E0930 14:41:25.104915 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="sg-core" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.104925 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="sg-core" Sep 30 14:41:25 crc kubenswrapper[4799]: E0930 14:41:25.104940 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="proxy-httpd" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.104948 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="proxy-httpd" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.105224 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="ceilometer-central-agent" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.105249 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="proxy-httpd" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.105263 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="sg-core" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.105282 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" containerName="ceilometer-notification-agent" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.108063 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.111799 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.111799 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.130975 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.185812 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phkh8\" (UniqueName: \"kubernetes.io/projected/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-kube-api-access-phkh8\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.185867 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-config-data\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.185895 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.185955 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-scripts\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.185979 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-run-httpd\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.186033 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.186056 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-log-httpd\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.287962 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phkh8\" (UniqueName: \"kubernetes.io/projected/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-kube-api-access-phkh8\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.288043 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-config-data\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.288091 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.288187 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-scripts\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.288227 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-run-httpd\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.288316 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.288354 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-log-httpd\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.288876 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-log-httpd\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.291098 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-run-httpd\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.293422 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-scripts\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.293487 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-config-data\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.294187 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.294638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.309753 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phkh8\" (UniqueName: \"kubernetes.io/projected/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-kube-api-access-phkh8\") pod \"ceilometer-0\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.426371 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.627961 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.746783 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da52c15f-1f9f-4485-99bc-1ea6b29c94c8","Type":"ContainerStarted","Data":"19e238ae7220945f702e3c0b3805c0451f01737dc325ad39c4590fd17de6f9a3"} Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.749530 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.771766 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.771746425 podStartE2EDuration="4.771746425s" podCreationTimestamp="2025-09-30 14:41:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:41:25.768348325 +0000 UTC m=+1307.851948752" watchObservedRunningTime="2025-09-30 14:41:25.771746425 +0000 UTC m=+1307.855346852" Sep 30 14:41:25 crc kubenswrapper[4799]: I0930 14:41:25.996753 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:26 crc kubenswrapper[4799]: I0930 14:41:26.541864 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fefc41cf-7c0f-43f9-a41b-01412bd697a6" path="/var/lib/kubelet/pods/fefc41cf-7c0f-43f9-a41b-01412bd697a6/volumes" Sep 30 14:41:26 crc kubenswrapper[4799]: I0930 14:41:26.757951 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerStarted","Data":"5de00bbe97cf469172075195c450f024df7df681afdd37910a805d93865e1da5"} Sep 30 14:41:27 crc kubenswrapper[4799]: I0930 14:41:27.528694 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:41:27 crc kubenswrapper[4799]: I0930 14:41:27.529048 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:41:27 crc kubenswrapper[4799]: I0930 14:41:27.684327 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:41:27 crc kubenswrapper[4799]: I0930 14:41:27.772287 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerStarted","Data":"340c7307e5b5dff04d684a0572197059c17fad428ebbe61f29af6ea85ee322ba"} Sep 30 14:41:27 crc kubenswrapper[4799]: I0930 14:41:27.772358 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerStarted","Data":"0d23ae0930d064eae413a83a3f99fe6bb517bde1e975b34eaa52ccac52a8fa5e"} Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.209609 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-s9qbm"] Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.211042 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-s9qbm" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.240801 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-s9qbm"] Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.270314 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmt5w\" (UniqueName: \"kubernetes.io/projected/99b988bd-7bdb-455e-a946-0586b3be1e7c-kube-api-access-lmt5w\") pod \"nova-api-db-create-s9qbm\" (UID: \"99b988bd-7bdb-455e-a946-0586b3be1e7c\") " pod="openstack/nova-api-db-create-s9qbm" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.348718 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-74vk7"] Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.350873 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-74vk7" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.372276 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmt5w\" (UniqueName: \"kubernetes.io/projected/99b988bd-7bdb-455e-a946-0586b3be1e7c-kube-api-access-lmt5w\") pod \"nova-api-db-create-s9qbm\" (UID: \"99b988bd-7bdb-455e-a946-0586b3be1e7c\") " pod="openstack/nova-api-db-create-s9qbm" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.399139 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmt5w\" (UniqueName: \"kubernetes.io/projected/99b988bd-7bdb-455e-a946-0586b3be1e7c-kube-api-access-lmt5w\") pod \"nova-api-db-create-s9qbm\" (UID: \"99b988bd-7bdb-455e-a946-0586b3be1e7c\") " pod="openstack/nova-api-db-create-s9qbm" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.399231 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-74vk7"] Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.474758 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-279sn\" (UniqueName: \"kubernetes.io/projected/fcfcac79-a568-4c8a-aff4-5533d18d954d-kube-api-access-279sn\") pod \"nova-cell0-db-create-74vk7\" (UID: \"fcfcac79-a568-4c8a-aff4-5533d18d954d\") " pod="openstack/nova-cell0-db-create-74vk7" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.526873 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-fpsgw"] Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.534495 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-s9qbm" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.535571 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fpsgw" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.545436 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fpsgw"] Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.577064 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s4lr\" (UniqueName: \"kubernetes.io/projected/d17315e1-62f6-4402-9b3e-90a88c4beaba-kube-api-access-4s4lr\") pod \"nova-cell1-db-create-fpsgw\" (UID: \"d17315e1-62f6-4402-9b3e-90a88c4beaba\") " pod="openstack/nova-cell1-db-create-fpsgw" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.577241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-279sn\" (UniqueName: \"kubernetes.io/projected/fcfcac79-a568-4c8a-aff4-5533d18d954d-kube-api-access-279sn\") pod \"nova-cell0-db-create-74vk7\" (UID: \"fcfcac79-a568-4c8a-aff4-5533d18d954d\") " pod="openstack/nova-cell0-db-create-74vk7" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.631455 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-279sn\" (UniqueName: \"kubernetes.io/projected/fcfcac79-a568-4c8a-aff4-5533d18d954d-kube-api-access-279sn\") pod \"nova-cell0-db-create-74vk7\" (UID: \"fcfcac79-a568-4c8a-aff4-5533d18d954d\") " pod="openstack/nova-cell0-db-create-74vk7" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.682791 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s4lr\" (UniqueName: \"kubernetes.io/projected/d17315e1-62f6-4402-9b3e-90a88c4beaba-kube-api-access-4s4lr\") pod \"nova-cell1-db-create-fpsgw\" (UID: \"d17315e1-62f6-4402-9b3e-90a88c4beaba\") " pod="openstack/nova-cell1-db-create-fpsgw" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.710978 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s4lr\" (UniqueName: \"kubernetes.io/projected/d17315e1-62f6-4402-9b3e-90a88c4beaba-kube-api-access-4s4lr\") pod \"nova-cell1-db-create-fpsgw\" (UID: \"d17315e1-62f6-4402-9b3e-90a88c4beaba\") " pod="openstack/nova-cell1-db-create-fpsgw" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.783037 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-74vk7" Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.789986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerStarted","Data":"eb21c0beb4a95b14e9dbb58f607006b1a3789357d9b7d72e8263292b6df2333e"} Sep 30 14:41:28 crc kubenswrapper[4799]: I0930 14:41:28.884640 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fpsgw" Sep 30 14:41:29 crc kubenswrapper[4799]: I0930 14:41:29.329329 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-s9qbm"] Sep 30 14:41:29 crc kubenswrapper[4799]: I0930 14:41:29.588060 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fpsgw"] Sep 30 14:41:29 crc kubenswrapper[4799]: I0930 14:41:29.649881 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:41:29 crc kubenswrapper[4799]: I0930 14:41:29.649945 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:41:29 crc kubenswrapper[4799]: I0930 14:41:29.693246 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-74vk7"] Sep 30 14:41:29 crc kubenswrapper[4799]: I0930 14:41:29.811580 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-74vk7" event={"ID":"fcfcac79-a568-4c8a-aff4-5533d18d954d","Type":"ContainerStarted","Data":"50e2a359fb47b761cb49e94305d45070efd9884a576adb3812c13de7d4b41f26"} Sep 30 14:41:29 crc kubenswrapper[4799]: I0930 14:41:29.822682 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fpsgw" event={"ID":"d17315e1-62f6-4402-9b3e-90a88c4beaba","Type":"ContainerStarted","Data":"78ab981ac68254655b7ad6c54f70560fab2dd3b90c061901d11d0c4816c98d6f"} Sep 30 14:41:29 crc kubenswrapper[4799]: I0930 14:41:29.833570 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-s9qbm" event={"ID":"99b988bd-7bdb-455e-a946-0586b3be1e7c","Type":"ContainerStarted","Data":"92af84b4f78f1b282351771a7de51bef086c8d5235c618a6023c04538b6cba90"} Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.850712 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-s9qbm" event={"ID":"99b988bd-7bdb-455e-a946-0586b3be1e7c","Type":"ContainerDied","Data":"58bcbfc7a3b37eee3dee8f9d16542b29635d111d0574dd06c3dfce1c3c268a5f"} Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.851374 4799 generic.go:334] "Generic (PLEG): container finished" podID="99b988bd-7bdb-455e-a946-0586b3be1e7c" containerID="58bcbfc7a3b37eee3dee8f9d16542b29635d111d0574dd06c3dfce1c3c268a5f" exitCode=0 Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.856819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerStarted","Data":"167326b9b3fc80993d6cf0c5bc23731fa5c5324f06a7d641869ba38eff8bbab5"} Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.857022 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.857067 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="ceilometer-central-agent" containerID="cri-o://0d23ae0930d064eae413a83a3f99fe6bb517bde1e975b34eaa52ccac52a8fa5e" gracePeriod=30 Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.857144 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="proxy-httpd" containerID="cri-o://167326b9b3fc80993d6cf0c5bc23731fa5c5324f06a7d641869ba38eff8bbab5" gracePeriod=30 Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.857110 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="ceilometer-notification-agent" containerID="cri-o://340c7307e5b5dff04d684a0572197059c17fad428ebbe61f29af6ea85ee322ba" gracePeriod=30 Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.857159 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="sg-core" containerID="cri-o://eb21c0beb4a95b14e9dbb58f607006b1a3789357d9b7d72e8263292b6df2333e" gracePeriod=30 Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.861956 4799 generic.go:334] "Generic (PLEG): container finished" podID="fcfcac79-a568-4c8a-aff4-5533d18d954d" containerID="45d49dee038e752e605be23615d61aafcd35140e6bc6e1ba7d730f3847ce766b" exitCode=0 Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.862070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-74vk7" event={"ID":"fcfcac79-a568-4c8a-aff4-5533d18d954d","Type":"ContainerDied","Data":"45d49dee038e752e605be23615d61aafcd35140e6bc6e1ba7d730f3847ce766b"} Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.867450 4799 generic.go:334] "Generic (PLEG): container finished" podID="d17315e1-62f6-4402-9b3e-90a88c4beaba" containerID="b6e5eb40df52f01f1e4de26c2853ea416973568133a2d4b7015558a8ab49bb5b" exitCode=0 Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.867728 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fpsgw" event={"ID":"d17315e1-62f6-4402-9b3e-90a88c4beaba","Type":"ContainerDied","Data":"b6e5eb40df52f01f1e4de26c2853ea416973568133a2d4b7015558a8ab49bb5b"} Sep 30 14:41:30 crc kubenswrapper[4799]: I0930 14:41:30.905880 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.551648455 podStartE2EDuration="5.905846645s" podCreationTimestamp="2025-09-30 14:41:25 +0000 UTC" firstStartedPulling="2025-09-30 14:41:25.999470772 +0000 UTC m=+1308.083071199" lastFinishedPulling="2025-09-30 14:41:30.353668952 +0000 UTC m=+1312.437269389" observedRunningTime="2025-09-30 14:41:30.891080603 +0000 UTC m=+1312.974681030" watchObservedRunningTime="2025-09-30 14:41:30.905846645 +0000 UTC m=+1312.989447072" Sep 30 14:41:31 crc kubenswrapper[4799]: I0930 14:41:31.882393 4799 generic.go:334] "Generic (PLEG): container finished" podID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerID="167326b9b3fc80993d6cf0c5bc23731fa5c5324f06a7d641869ba38eff8bbab5" exitCode=0 Sep 30 14:41:31 crc kubenswrapper[4799]: I0930 14:41:31.882698 4799 generic.go:334] "Generic (PLEG): container finished" podID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerID="eb21c0beb4a95b14e9dbb58f607006b1a3789357d9b7d72e8263292b6df2333e" exitCode=2 Sep 30 14:41:31 crc kubenswrapper[4799]: I0930 14:41:31.882713 4799 generic.go:334] "Generic (PLEG): container finished" podID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerID="340c7307e5b5dff04d684a0572197059c17fad428ebbe61f29af6ea85ee322ba" exitCode=0 Sep 30 14:41:31 crc kubenswrapper[4799]: I0930 14:41:31.882476 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerDied","Data":"167326b9b3fc80993d6cf0c5bc23731fa5c5324f06a7d641869ba38eff8bbab5"} Sep 30 14:41:31 crc kubenswrapper[4799]: I0930 14:41:31.882922 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerDied","Data":"eb21c0beb4a95b14e9dbb58f607006b1a3789357d9b7d72e8263292b6df2333e"} Sep 30 14:41:31 crc kubenswrapper[4799]: I0930 14:41:31.882943 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerDied","Data":"340c7307e5b5dff04d684a0572197059c17fad428ebbe61f29af6ea85ee322ba"} Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.387771 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fpsgw" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.530219 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s4lr\" (UniqueName: \"kubernetes.io/projected/d17315e1-62f6-4402-9b3e-90a88c4beaba-kube-api-access-4s4lr\") pod \"d17315e1-62f6-4402-9b3e-90a88c4beaba\" (UID: \"d17315e1-62f6-4402-9b3e-90a88c4beaba\") " Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.542290 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17315e1-62f6-4402-9b3e-90a88c4beaba-kube-api-access-4s4lr" (OuterVolumeSpecName: "kube-api-access-4s4lr") pod "d17315e1-62f6-4402-9b3e-90a88c4beaba" (UID: "d17315e1-62f6-4402-9b3e-90a88c4beaba"). InnerVolumeSpecName "kube-api-access-4s4lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.639438 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s4lr\" (UniqueName: \"kubernetes.io/projected/d17315e1-62f6-4402-9b3e-90a88c4beaba-kube-api-access-4s4lr\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.789974 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-s9qbm" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.841615 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-74vk7" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.926822 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-74vk7" event={"ID":"fcfcac79-a568-4c8a-aff4-5533d18d954d","Type":"ContainerDied","Data":"50e2a359fb47b761cb49e94305d45070efd9884a576adb3812c13de7d4b41f26"} Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.926895 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50e2a359fb47b761cb49e94305d45070efd9884a576adb3812c13de7d4b41f26" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.926992 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-74vk7" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.932733 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fpsgw" event={"ID":"d17315e1-62f6-4402-9b3e-90a88c4beaba","Type":"ContainerDied","Data":"78ab981ac68254655b7ad6c54f70560fab2dd3b90c061901d11d0c4816c98d6f"} Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.932788 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78ab981ac68254655b7ad6c54f70560fab2dd3b90c061901d11d0c4816c98d6f" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.932921 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fpsgw" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.939418 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-s9qbm" event={"ID":"99b988bd-7bdb-455e-a946-0586b3be1e7c","Type":"ContainerDied","Data":"92af84b4f78f1b282351771a7de51bef086c8d5235c618a6023c04538b6cba90"} Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.939483 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92af84b4f78f1b282351771a7de51bef086c8d5235c618a6023c04538b6cba90" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.939581 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-s9qbm" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.947567 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmt5w\" (UniqueName: \"kubernetes.io/projected/99b988bd-7bdb-455e-a946-0586b3be1e7c-kube-api-access-lmt5w\") pod \"99b988bd-7bdb-455e-a946-0586b3be1e7c\" (UID: \"99b988bd-7bdb-455e-a946-0586b3be1e7c\") " Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.947823 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279sn\" (UniqueName: \"kubernetes.io/projected/fcfcac79-a568-4c8a-aff4-5533d18d954d-kube-api-access-279sn\") pod \"fcfcac79-a568-4c8a-aff4-5533d18d954d\" (UID: \"fcfcac79-a568-4c8a-aff4-5533d18d954d\") " Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.952971 4799 generic.go:334] "Generic (PLEG): container finished" podID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerID="0d23ae0930d064eae413a83a3f99fe6bb517bde1e975b34eaa52ccac52a8fa5e" exitCode=0 Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.953021 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerDied","Data":"0d23ae0930d064eae413a83a3f99fe6bb517bde1e975b34eaa52ccac52a8fa5e"} Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.973006 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b988bd-7bdb-455e-a946-0586b3be1e7c-kube-api-access-lmt5w" (OuterVolumeSpecName: "kube-api-access-lmt5w") pod "99b988bd-7bdb-455e-a946-0586b3be1e7c" (UID: "99b988bd-7bdb-455e-a946-0586b3be1e7c"). InnerVolumeSpecName "kube-api-access-lmt5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:32 crc kubenswrapper[4799]: I0930 14:41:32.975876 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcfcac79-a568-4c8a-aff4-5533d18d954d-kube-api-access-279sn" (OuterVolumeSpecName: "kube-api-access-279sn") pod "fcfcac79-a568-4c8a-aff4-5533d18d954d" (UID: "fcfcac79-a568-4c8a-aff4-5533d18d954d"). InnerVolumeSpecName "kube-api-access-279sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.050476 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279sn\" (UniqueName: \"kubernetes.io/projected/fcfcac79-a568-4c8a-aff4-5533d18d954d-kube-api-access-279sn\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.050517 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmt5w\" (UniqueName: \"kubernetes.io/projected/99b988bd-7bdb-455e-a946-0586b3be1e7c-kube-api-access-lmt5w\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.248244 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.356704 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-scripts\") pod \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.356914 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-run-httpd\") pod \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.356964 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-combined-ca-bundle\") pod \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.357022 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phkh8\" (UniqueName: \"kubernetes.io/projected/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-kube-api-access-phkh8\") pod \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.357139 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-sg-core-conf-yaml\") pod \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.357200 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-log-httpd\") pod \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.357251 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-config-data\") pod \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\" (UID: \"6a478fc5-4f80-463e-a9a4-fed8d6030bf2\") " Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.363058 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6a478fc5-4f80-463e-a9a4-fed8d6030bf2" (UID: "6a478fc5-4f80-463e-a9a4-fed8d6030bf2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.363102 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6a478fc5-4f80-463e-a9a4-fed8d6030bf2" (UID: "6a478fc5-4f80-463e-a9a4-fed8d6030bf2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.393990 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-scripts" (OuterVolumeSpecName: "scripts") pod "6a478fc5-4f80-463e-a9a4-fed8d6030bf2" (UID: "6a478fc5-4f80-463e-a9a4-fed8d6030bf2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.394333 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-kube-api-access-phkh8" (OuterVolumeSpecName: "kube-api-access-phkh8") pod "6a478fc5-4f80-463e-a9a4-fed8d6030bf2" (UID: "6a478fc5-4f80-463e-a9a4-fed8d6030bf2"). InnerVolumeSpecName "kube-api-access-phkh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.439842 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6a478fc5-4f80-463e-a9a4-fed8d6030bf2" (UID: "6a478fc5-4f80-463e-a9a4-fed8d6030bf2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.461168 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phkh8\" (UniqueName: \"kubernetes.io/projected/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-kube-api-access-phkh8\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.461494 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.461572 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.461660 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.461742 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.531932 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-config-data" (OuterVolumeSpecName: "config-data") pod "6a478fc5-4f80-463e-a9a4-fed8d6030bf2" (UID: "6a478fc5-4f80-463e-a9a4-fed8d6030bf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.532088 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a478fc5-4f80-463e-a9a4-fed8d6030bf2" (UID: "6a478fc5-4f80-463e-a9a4-fed8d6030bf2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.564008 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.564047 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a478fc5-4f80-463e-a9a4-fed8d6030bf2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:33 crc kubenswrapper[4799]: E0930 14:41:33.582133 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99b988bd_7bdb_455e_a946_0586b3be1e7c.slice/crio-92af84b4f78f1b282351771a7de51bef086c8d5235c618a6023c04538b6cba90\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcfcac79_a568_4c8a_aff4_5533d18d954d.slice/crio-50e2a359fb47b761cb49e94305d45070efd9884a576adb3812c13de7d4b41f26\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99b988bd_7bdb_455e_a946_0586b3be1e7c.slice\": RecentStats: unable to find data in memory cache]" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.970210 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6a478fc5-4f80-463e-a9a4-fed8d6030bf2","Type":"ContainerDied","Data":"5de00bbe97cf469172075195c450f024df7df681afdd37910a805d93865e1da5"} Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.970298 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:33 crc kubenswrapper[4799]: I0930 14:41:33.970600 4799 scope.go:117] "RemoveContainer" containerID="167326b9b3fc80993d6cf0c5bc23731fa5c5324f06a7d641869ba38eff8bbab5" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.012366 4799 scope.go:117] "RemoveContainer" containerID="eb21c0beb4a95b14e9dbb58f607006b1a3789357d9b7d72e8263292b6df2333e" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.024396 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.051358 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.071556 4799 scope.go:117] "RemoveContainer" containerID="340c7307e5b5dff04d684a0572197059c17fad428ebbe61f29af6ea85ee322ba" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.096276 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:34 crc kubenswrapper[4799]: E0930 14:41:34.098917 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="ceilometer-central-agent" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.098970 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="ceilometer-central-agent" Sep 30 14:41:34 crc kubenswrapper[4799]: E0930 14:41:34.099010 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b988bd-7bdb-455e-a946-0586b3be1e7c" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099020 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b988bd-7bdb-455e-a946-0586b3be1e7c" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: E0930 14:41:34.099035 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="sg-core" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099043 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="sg-core" Sep 30 14:41:34 crc kubenswrapper[4799]: E0930 14:41:34.099051 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcfcac79-a568-4c8a-aff4-5533d18d954d" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099276 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcfcac79-a568-4c8a-aff4-5533d18d954d" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: E0930 14:41:34.099303 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="ceilometer-notification-agent" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099311 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="ceilometer-notification-agent" Sep 30 14:41:34 crc kubenswrapper[4799]: E0930 14:41:34.099339 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17315e1-62f6-4402-9b3e-90a88c4beaba" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099346 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17315e1-62f6-4402-9b3e-90a88c4beaba" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: E0930 14:41:34.099371 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="proxy-httpd" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099377 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="proxy-httpd" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099560 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17315e1-62f6-4402-9b3e-90a88c4beaba" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099571 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="proxy-httpd" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099581 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="ceilometer-central-agent" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099595 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="sg-core" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099613 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" containerName="ceilometer-notification-agent" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099623 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcfcac79-a568-4c8a-aff4-5533d18d954d" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.099634 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b988bd-7bdb-455e-a946-0586b3be1e7c" containerName="mariadb-database-create" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.101797 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.104627 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.107920 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.143756 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.150209 4799 scope.go:117] "RemoveContainer" containerID="0d23ae0930d064eae413a83a3f99fe6bb517bde1e975b34eaa52ccac52a8fa5e" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.178815 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.178902 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.178944 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-log-httpd\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.179001 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-run-httpd\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.179131 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hntlz\" (UniqueName: \"kubernetes.io/projected/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-kube-api-access-hntlz\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.179177 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-config-data\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.179195 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-scripts\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.281525 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.281596 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.281667 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-log-httpd\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.281729 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-run-httpd\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.281775 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hntlz\" (UniqueName: \"kubernetes.io/projected/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-kube-api-access-hntlz\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.281809 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-config-data\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.281825 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-scripts\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.282532 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-log-httpd\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.282858 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-run-httpd\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.286326 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-scripts\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.286429 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.290448 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.297119 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-config-data\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.312552 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hntlz\" (UniqueName: \"kubernetes.io/projected/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-kube-api-access-hntlz\") pod \"ceilometer-0\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.445817 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:34 crc kubenswrapper[4799]: I0930 14:41:34.529444 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a478fc5-4f80-463e-a9a4-fed8d6030bf2" path="/var/lib/kubelet/pods/6a478fc5-4f80-463e-a9a4-fed8d6030bf2/volumes" Sep 30 14:41:35 crc kubenswrapper[4799]: I0930 14:41:35.090284 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:41:35 crc kubenswrapper[4799]: I0930 14:41:35.114470 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:35 crc kubenswrapper[4799]: I0930 14:41:35.999064 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerStarted","Data":"0188357c541e2fbfd812baa0b2f877fe91163e82b4fa5874b7acae76cd82cf27"} Sep 30 14:41:36 crc kubenswrapper[4799]: I0930 14:41:36.090849 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="da52c15f-1f9f-4485-99bc-1ea6b29c94c8" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.171:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:41:37 crc kubenswrapper[4799]: I0930 14:41:37.012209 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerStarted","Data":"ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3"} Sep 30 14:41:37 crc kubenswrapper[4799]: I0930 14:41:37.088970 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="da52c15f-1f9f-4485-99bc-1ea6b29c94c8" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.171:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:41:37 crc kubenswrapper[4799]: I0930 14:41:37.531326 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:41:37 crc kubenswrapper[4799]: I0930 14:41:37.684132 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.029029 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerStarted","Data":"85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec"} Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.461990 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-2e33-account-create-t58d8"] Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.463577 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2e33-account-create-t58d8" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.470185 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.488383 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2e33-account-create-t58d8"] Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.601425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s9j9\" (UniqueName: \"kubernetes.io/projected/d70a7325-993b-4f20-a06c-57cdbeb7d557-kube-api-access-9s9j9\") pod \"nova-api-2e33-account-create-t58d8\" (UID: \"d70a7325-993b-4f20-a06c-57cdbeb7d557\") " pod="openstack/nova-api-2e33-account-create-t58d8" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.705854 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s9j9\" (UniqueName: \"kubernetes.io/projected/d70a7325-993b-4f20-a06c-57cdbeb7d557-kube-api-access-9s9j9\") pod \"nova-api-2e33-account-create-t58d8\" (UID: \"d70a7325-993b-4f20-a06c-57cdbeb7d557\") " pod="openstack/nova-api-2e33-account-create-t58d8" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.760931 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8bff-account-create-6pft6"] Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.763498 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bff-account-create-6pft6" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.769358 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s9j9\" (UniqueName: \"kubernetes.io/projected/d70a7325-993b-4f20-a06c-57cdbeb7d557-kube-api-access-9s9j9\") pod \"nova-api-2e33-account-create-t58d8\" (UID: \"d70a7325-993b-4f20-a06c-57cdbeb7d557\") " pod="openstack/nova-api-2e33-account-create-t58d8" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.779669 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.784244 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2e33-account-create-t58d8" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.785017 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8bff-account-create-6pft6"] Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.812925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g6td\" (UniqueName: \"kubernetes.io/projected/d43c94bd-d76c-4680-bf8d-e2f02fd2279c-kube-api-access-6g6td\") pod \"nova-cell1-8bff-account-create-6pft6\" (UID: \"d43c94bd-d76c-4680-bf8d-e2f02fd2279c\") " pod="openstack/nova-cell1-8bff-account-create-6pft6" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.914903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g6td\" (UniqueName: \"kubernetes.io/projected/d43c94bd-d76c-4680-bf8d-e2f02fd2279c-kube-api-access-6g6td\") pod \"nova-cell1-8bff-account-create-6pft6\" (UID: \"d43c94bd-d76c-4680-bf8d-e2f02fd2279c\") " pod="openstack/nova-cell1-8bff-account-create-6pft6" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.943232 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g6td\" (UniqueName: \"kubernetes.io/projected/d43c94bd-d76c-4680-bf8d-e2f02fd2279c-kube-api-access-6g6td\") pod \"nova-cell1-8bff-account-create-6pft6\" (UID: \"d43c94bd-d76c-4680-bf8d-e2f02fd2279c\") " pod="openstack/nova-cell1-8bff-account-create-6pft6" Sep 30 14:41:38 crc kubenswrapper[4799]: I0930 14:41:38.944129 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bff-account-create-6pft6" Sep 30 14:41:39 crc kubenswrapper[4799]: I0930 14:41:39.127079 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerStarted","Data":"64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366"} Sep 30 14:41:39 crc kubenswrapper[4799]: I0930 14:41:39.423370 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2e33-account-create-t58d8"] Sep 30 14:41:40 crc kubenswrapper[4799]: I0930 14:41:40.030366 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8bff-account-create-6pft6"] Sep 30 14:41:40 crc kubenswrapper[4799]: I0930 14:41:40.156382 4799 generic.go:334] "Generic (PLEG): container finished" podID="d70a7325-993b-4f20-a06c-57cdbeb7d557" containerID="f16a947cdc5b5f1a40d76a7948f221b7b454b1a6dc7cde5ea7f0ece434af8582" exitCode=0 Sep 30 14:41:40 crc kubenswrapper[4799]: I0930 14:41:40.156474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2e33-account-create-t58d8" event={"ID":"d70a7325-993b-4f20-a06c-57cdbeb7d557","Type":"ContainerDied","Data":"f16a947cdc5b5f1a40d76a7948f221b7b454b1a6dc7cde5ea7f0ece434af8582"} Sep 30 14:41:40 crc kubenswrapper[4799]: I0930 14:41:40.156511 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2e33-account-create-t58d8" event={"ID":"d70a7325-993b-4f20-a06c-57cdbeb7d557","Type":"ContainerStarted","Data":"191d1a7ea8b797dcacd8165b766e1d5c93ce425bdff397331189292b1da8b826"} Sep 30 14:41:40 crc kubenswrapper[4799]: I0930 14:41:40.162996 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bff-account-create-6pft6" event={"ID":"d43c94bd-d76c-4680-bf8d-e2f02fd2279c","Type":"ContainerStarted","Data":"c6d2ac032320e4ef39124b8a38ea9bc11f3f44635f5e63ae3117270aa3216b4e"} Sep 30 14:41:40 crc kubenswrapper[4799]: I0930 14:41:40.167936 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerStarted","Data":"251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07"} Sep 30 14:41:40 crc kubenswrapper[4799]: I0930 14:41:40.169096 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:41:40 crc kubenswrapper[4799]: I0930 14:41:40.218690 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.821391867 podStartE2EDuration="6.218667435s" podCreationTimestamp="2025-09-30 14:41:34 +0000 UTC" firstStartedPulling="2025-09-30 14:41:35.089909403 +0000 UTC m=+1317.173509830" lastFinishedPulling="2025-09-30 14:41:39.487184971 +0000 UTC m=+1321.570785398" observedRunningTime="2025-09-30 14:41:40.211565538 +0000 UTC m=+1322.295165995" watchObservedRunningTime="2025-09-30 14:41:40.218667435 +0000 UTC m=+1322.302267862" Sep 30 14:41:41 crc kubenswrapper[4799]: I0930 14:41:41.096979 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="da52c15f-1f9f-4485-99bc-1ea6b29c94c8" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.171:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:41:41 crc kubenswrapper[4799]: I0930 14:41:41.186714 4799 generic.go:334] "Generic (PLEG): container finished" podID="d43c94bd-d76c-4680-bf8d-e2f02fd2279c" containerID="e5b4375d1706fe86bf8d15415458e779a1ce85e7b57531385483f6b52d5e568a" exitCode=0 Sep 30 14:41:41 crc kubenswrapper[4799]: I0930 14:41:41.186781 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bff-account-create-6pft6" event={"ID":"d43c94bd-d76c-4680-bf8d-e2f02fd2279c","Type":"ContainerDied","Data":"e5b4375d1706fe86bf8d15415458e779a1ce85e7b57531385483f6b52d5e568a"} Sep 30 14:41:41 crc kubenswrapper[4799]: I0930 14:41:41.770708 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2e33-account-create-t58d8" Sep 30 14:41:41 crc kubenswrapper[4799]: I0930 14:41:41.826539 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s9j9\" (UniqueName: \"kubernetes.io/projected/d70a7325-993b-4f20-a06c-57cdbeb7d557-kube-api-access-9s9j9\") pod \"d70a7325-993b-4f20-a06c-57cdbeb7d557\" (UID: \"d70a7325-993b-4f20-a06c-57cdbeb7d557\") " Sep 30 14:41:41 crc kubenswrapper[4799]: I0930 14:41:41.850066 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d70a7325-993b-4f20-a06c-57cdbeb7d557-kube-api-access-9s9j9" (OuterVolumeSpecName: "kube-api-access-9s9j9") pod "d70a7325-993b-4f20-a06c-57cdbeb7d557" (UID: "d70a7325-993b-4f20-a06c-57cdbeb7d557"). InnerVolumeSpecName "kube-api-access-9s9j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:41 crc kubenswrapper[4799]: I0930 14:41:41.928860 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s9j9\" (UniqueName: \"kubernetes.io/projected/d70a7325-993b-4f20-a06c-57cdbeb7d557-kube-api-access-9s9j9\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:42 crc kubenswrapper[4799]: I0930 14:41:42.093895 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="da52c15f-1f9f-4485-99bc-1ea6b29c94c8" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.171:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:41:42 crc kubenswrapper[4799]: I0930 14:41:42.109743 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Sep 30 14:41:42 crc kubenswrapper[4799]: I0930 14:41:42.271231 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2e33-account-create-t58d8" Sep 30 14:41:42 crc kubenswrapper[4799]: I0930 14:41:42.275816 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2e33-account-create-t58d8" event={"ID":"d70a7325-993b-4f20-a06c-57cdbeb7d557","Type":"ContainerDied","Data":"191d1a7ea8b797dcacd8165b766e1d5c93ce425bdff397331189292b1da8b826"} Sep 30 14:41:42 crc kubenswrapper[4799]: I0930 14:41:42.275888 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="191d1a7ea8b797dcacd8165b766e1d5c93ce425bdff397331189292b1da8b826" Sep 30 14:41:42 crc kubenswrapper[4799]: I0930 14:41:42.941216 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bff-account-create-6pft6" Sep 30 14:41:42 crc kubenswrapper[4799]: I0930 14:41:42.973117 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6td\" (UniqueName: \"kubernetes.io/projected/d43c94bd-d76c-4680-bf8d-e2f02fd2279c-kube-api-access-6g6td\") pod \"d43c94bd-d76c-4680-bf8d-e2f02fd2279c\" (UID: \"d43c94bd-d76c-4680-bf8d-e2f02fd2279c\") " Sep 30 14:41:42 crc kubenswrapper[4799]: I0930 14:41:42.979243 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d43c94bd-d76c-4680-bf8d-e2f02fd2279c-kube-api-access-6g6td" (OuterVolumeSpecName: "kube-api-access-6g6td") pod "d43c94bd-d76c-4680-bf8d-e2f02fd2279c" (UID: "d43c94bd-d76c-4680-bf8d-e2f02fd2279c"). InnerVolumeSpecName "kube-api-access-6g6td". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:43 crc kubenswrapper[4799]: I0930 14:41:43.075692 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6td\" (UniqueName: \"kubernetes.io/projected/d43c94bd-d76c-4680-bf8d-e2f02fd2279c-kube-api-access-6g6td\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:43 crc kubenswrapper[4799]: I0930 14:41:43.285493 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bff-account-create-6pft6" event={"ID":"d43c94bd-d76c-4680-bf8d-e2f02fd2279c","Type":"ContainerDied","Data":"c6d2ac032320e4ef39124b8a38ea9bc11f3f44635f5e63ae3117270aa3216b4e"} Sep 30 14:41:43 crc kubenswrapper[4799]: I0930 14:41:43.285872 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6d2ac032320e4ef39124b8a38ea9bc11f3f44635f5e63ae3117270aa3216b4e" Sep 30 14:41:43 crc kubenswrapper[4799]: I0930 14:41:43.285956 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bff-account-create-6pft6" Sep 30 14:41:45 crc kubenswrapper[4799]: I0930 14:41:45.427182 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:45 crc kubenswrapper[4799]: I0930 14:41:45.427570 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="ceilometer-central-agent" containerID="cri-o://ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3" gracePeriod=30 Sep 30 14:41:45 crc kubenswrapper[4799]: I0930 14:41:45.428252 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="proxy-httpd" containerID="cri-o://251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07" gracePeriod=30 Sep 30 14:41:45 crc kubenswrapper[4799]: I0930 14:41:45.428345 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="sg-core" containerID="cri-o://64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366" gracePeriod=30 Sep 30 14:41:45 crc kubenswrapper[4799]: I0930 14:41:45.428422 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="ceilometer-notification-agent" containerID="cri-o://85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec" gracePeriod=30 Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.322828 4799 generic.go:334] "Generic (PLEG): container finished" podID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerID="251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07" exitCode=0 Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.323164 4799 generic.go:334] "Generic (PLEG): container finished" podID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerID="64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366" exitCode=2 Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.323181 4799 generic.go:334] "Generic (PLEG): container finished" podID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerID="85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec" exitCode=0 Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.324355 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerDied","Data":"251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07"} Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.324399 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerDied","Data":"64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366"} Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.324412 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerDied","Data":"85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec"} Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.832526 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.868341 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-config-data\") pod \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.868403 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-log-httpd\") pod \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.868453 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-scripts\") pod \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.868501 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hntlz\" (UniqueName: \"kubernetes.io/projected/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-kube-api-access-hntlz\") pod \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.868591 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-run-httpd\") pod \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.868617 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-combined-ca-bundle\") pod \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.868793 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-sg-core-conf-yaml\") pod \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\" (UID: \"bff4c9a1-45af-41ef-ae6c-6183f5727b8c\") " Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.869728 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bff4c9a1-45af-41ef-ae6c-6183f5727b8c" (UID: "bff4c9a1-45af-41ef-ae6c-6183f5727b8c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.869945 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bff4c9a1-45af-41ef-ae6c-6183f5727b8c" (UID: "bff4c9a1-45af-41ef-ae6c-6183f5727b8c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.879285 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-scripts" (OuterVolumeSpecName: "scripts") pod "bff4c9a1-45af-41ef-ae6c-6183f5727b8c" (UID: "bff4c9a1-45af-41ef-ae6c-6183f5727b8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.884087 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-kube-api-access-hntlz" (OuterVolumeSpecName: "kube-api-access-hntlz") pod "bff4c9a1-45af-41ef-ae6c-6183f5727b8c" (UID: "bff4c9a1-45af-41ef-ae6c-6183f5727b8c"). InnerVolumeSpecName "kube-api-access-hntlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.959899 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bff4c9a1-45af-41ef-ae6c-6183f5727b8c" (UID: "bff4c9a1-45af-41ef-ae6c-6183f5727b8c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.974594 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.974684 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.974697 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.974708 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:46 crc kubenswrapper[4799]: I0930 14:41:46.974724 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hntlz\" (UniqueName: \"kubernetes.io/projected/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-kube-api-access-hntlz\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.070676 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bff4c9a1-45af-41ef-ae6c-6183f5727b8c" (UID: "bff4c9a1-45af-41ef-ae6c-6183f5727b8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.086372 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.135876 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-config-data" (OuterVolumeSpecName: "config-data") pod "bff4c9a1-45af-41ef-ae6c-6183f5727b8c" (UID: "bff4c9a1-45af-41ef-ae6c-6183f5727b8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.188813 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff4c9a1-45af-41ef-ae6c-6183f5727b8c-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.341407 4799 generic.go:334] "Generic (PLEG): container finished" podID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerID="ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3" exitCode=0 Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.341475 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerDied","Data":"ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3"} Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.341807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bff4c9a1-45af-41ef-ae6c-6183f5727b8c","Type":"ContainerDied","Data":"0188357c541e2fbfd812baa0b2f877fe91163e82b4fa5874b7acae76cd82cf27"} Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.341833 4799 scope.go:117] "RemoveContainer" containerID="251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.341536 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.373464 4799 scope.go:117] "RemoveContainer" containerID="64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.398616 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.410516 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.425274 4799 scope.go:117] "RemoveContainer" containerID="85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.452476 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.453212 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d43c94bd-d76c-4680-bf8d-e2f02fd2279c" containerName="mariadb-account-create" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453235 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d43c94bd-d76c-4680-bf8d-e2f02fd2279c" containerName="mariadb-account-create" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.453258 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="ceilometer-notification-agent" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453267 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="ceilometer-notification-agent" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.453290 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d70a7325-993b-4f20-a06c-57cdbeb7d557" containerName="mariadb-account-create" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453299 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d70a7325-993b-4f20-a06c-57cdbeb7d557" containerName="mariadb-account-create" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.453314 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="sg-core" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453322 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="sg-core" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.453342 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="ceilometer-central-agent" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453351 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="ceilometer-central-agent" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.453369 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="proxy-httpd" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453376 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="proxy-httpd" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453611 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="sg-core" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453637 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="ceilometer-notification-agent" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453758 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="ceilometer-central-agent" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453776 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d43c94bd-d76c-4680-bf8d-e2f02fd2279c" containerName="mariadb-account-create" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453799 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" containerName="proxy-httpd" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.453814 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d70a7325-993b-4f20-a06c-57cdbeb7d557" containerName="mariadb-account-create" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.458240 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.470109 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.470389 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.477422 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.479830 4799 scope.go:117] "RemoveContainer" containerID="ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.497628 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.497780 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-config-data\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.497811 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.497848 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-run-httpd\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.497868 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-scripts\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.497955 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-log-httpd\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.498078 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f64j\" (UniqueName: \"kubernetes.io/projected/ba1bdf4c-e42d-4da6-af94-9c6f23569426-kube-api-access-2f64j\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.537457 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.596624 4799 scope.go:117] "RemoveContainer" containerID="251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.599425 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-log-httpd\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.599488 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f64j\" (UniqueName: \"kubernetes.io/projected/ba1bdf4c-e42d-4da6-af94-9c6f23569426-kube-api-access-2f64j\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.599569 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.599631 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-config-data\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.599695 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.599733 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-scripts\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.599752 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-run-httpd\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.600324 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-run-httpd\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.600575 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-log-httpd\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.601513 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07\": container with ID starting with 251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07 not found: ID does not exist" containerID="251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.601554 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07"} err="failed to get container status \"251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07\": rpc error: code = NotFound desc = could not find container \"251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07\": container with ID starting with 251228c776219971af6a052df74190a5be16b632275fe16a76bf766fc4466a07 not found: ID does not exist" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.601600 4799 scope.go:117] "RemoveContainer" containerID="64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.606870 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366\": container with ID starting with 64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366 not found: ID does not exist" containerID="64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.606929 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366"} err="failed to get container status \"64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366\": rpc error: code = NotFound desc = could not find container \"64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366\": container with ID starting with 64f66854c95eb3a2979bde9245cf6bac1371df78dd45de5b8fac2ed968dfb366 not found: ID does not exist" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.606966 4799 scope.go:117] "RemoveContainer" containerID="85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.607824 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.607965 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec\": container with ID starting with 85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec not found: ID does not exist" containerID="85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.607993 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec"} err="failed to get container status \"85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec\": rpc error: code = NotFound desc = could not find container \"85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec\": container with ID starting with 85275104089a991aaf571b53e57bd5a0e2ca5dee561813181997ed3cf57ef2ec not found: ID does not exist" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.608016 4799 scope.go:117] "RemoveContainer" containerID="ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.609276 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-scripts\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: E0930 14:41:47.613842 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3\": container with ID starting with ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3 not found: ID does not exist" containerID="ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.613903 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3"} err="failed to get container status \"ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3\": rpc error: code = NotFound desc = could not find container \"ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3\": container with ID starting with ab41c211cb81890ee228f167f0a08401913c19b83f2242265bec955dcb3de1e3 not found: ID does not exist" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.618101 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.623747 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-config-data\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.628431 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f64j\" (UniqueName: \"kubernetes.io/projected/ba1bdf4c-e42d-4da6-af94-9c6f23569426-kube-api-access-2f64j\") pod \"ceilometer-0\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " pod="openstack/ceilometer-0" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.682828 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.682936 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.684119 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"2aa2de6ee9c44af7d31b1105ea4177bb893946376162ccd5d747ef4d0d8305e3"} pod="openstack/horizon-7f7c888686-pwv2q" containerMessage="Container horizon failed startup probe, will be restarted" Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.684178 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" containerID="cri-o://2aa2de6ee9c44af7d31b1105ea4177bb893946376162ccd5d747ef4d0d8305e3" gracePeriod=30 Sep 30 14:41:47 crc kubenswrapper[4799]: I0930 14:41:47.786263 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.340381 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.356744 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerStarted","Data":"5fa943f02c2c00583aff101e622149e944220948f2297a4e76f8d9443bb163ff"} Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.515862 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bff4c9a1-45af-41ef-ae6c-6183f5727b8c" path="/var/lib/kubelet/pods/bff4c9a1-45af-41ef-ae6c-6183f5727b8c/volumes" Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.554183 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-2670-account-create-z7z29"] Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.555776 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2670-account-create-z7z29" Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.561809 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.628393 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9h5w\" (UniqueName: \"kubernetes.io/projected/08dca185-ba7a-439d-96ee-2e5aba753dc5-kube-api-access-k9h5w\") pod \"nova-cell0-2670-account-create-z7z29\" (UID: \"08dca185-ba7a-439d-96ee-2e5aba753dc5\") " pod="openstack/nova-cell0-2670-account-create-z7z29" Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.633853 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2670-account-create-z7z29"] Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.731444 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9h5w\" (UniqueName: \"kubernetes.io/projected/08dca185-ba7a-439d-96ee-2e5aba753dc5-kube-api-access-k9h5w\") pod \"nova-cell0-2670-account-create-z7z29\" (UID: \"08dca185-ba7a-439d-96ee-2e5aba753dc5\") " pod="openstack/nova-cell0-2670-account-create-z7z29" Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.756902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9h5w\" (UniqueName: \"kubernetes.io/projected/08dca185-ba7a-439d-96ee-2e5aba753dc5-kube-api-access-k9h5w\") pod \"nova-cell0-2670-account-create-z7z29\" (UID: \"08dca185-ba7a-439d-96ee-2e5aba753dc5\") " pod="openstack/nova-cell0-2670-account-create-z7z29" Sep 30 14:41:48 crc kubenswrapper[4799]: I0930 14:41:48.905032 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2670-account-create-z7z29" Sep 30 14:41:49 crc kubenswrapper[4799]: I0930 14:41:49.518695 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2670-account-create-z7z29"] Sep 30 14:41:50 crc kubenswrapper[4799]: I0930 14:41:50.386200 4799 generic.go:334] "Generic (PLEG): container finished" podID="08dca185-ba7a-439d-96ee-2e5aba753dc5" containerID="2b33610392cc6954114682f78227200f563c71da7b917eb4d9dd69af047008b0" exitCode=0 Sep 30 14:41:50 crc kubenswrapper[4799]: I0930 14:41:50.386299 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2670-account-create-z7z29" event={"ID":"08dca185-ba7a-439d-96ee-2e5aba753dc5","Type":"ContainerDied","Data":"2b33610392cc6954114682f78227200f563c71da7b917eb4d9dd69af047008b0"} Sep 30 14:41:50 crc kubenswrapper[4799]: I0930 14:41:50.386673 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2670-account-create-z7z29" event={"ID":"08dca185-ba7a-439d-96ee-2e5aba753dc5","Type":"ContainerStarted","Data":"79cd3d4d913c03471789adb2ca8bc6de40ce99da307de3e8c0895e89fd93e400"} Sep 30 14:41:50 crc kubenswrapper[4799]: I0930 14:41:50.389133 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerStarted","Data":"32fa7121b9a6200b71434ee14f58f183fde97f26fa021511cdaf7794213b491d"} Sep 30 14:41:51 crc kubenswrapper[4799]: I0930 14:41:51.402731 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerStarted","Data":"713617e71c53f603a2a075b33b4a2d4f713d8b1d2377b5c739ca098e76d3c79f"} Sep 30 14:41:51 crc kubenswrapper[4799]: I0930 14:41:51.403070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerStarted","Data":"fa48171e870ac0e6c827755093aa34ec1156ebefb32dad88d1cafdc5f78e10a0"} Sep 30 14:41:51 crc kubenswrapper[4799]: I0930 14:41:51.440816 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:51 crc kubenswrapper[4799]: I0930 14:41:51.851632 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2670-account-create-z7z29" Sep 30 14:41:51 crc kubenswrapper[4799]: I0930 14:41:51.903281 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9h5w\" (UniqueName: \"kubernetes.io/projected/08dca185-ba7a-439d-96ee-2e5aba753dc5-kube-api-access-k9h5w\") pod \"08dca185-ba7a-439d-96ee-2e5aba753dc5\" (UID: \"08dca185-ba7a-439d-96ee-2e5aba753dc5\") " Sep 30 14:41:51 crc kubenswrapper[4799]: I0930 14:41:51.911690 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08dca185-ba7a-439d-96ee-2e5aba753dc5-kube-api-access-k9h5w" (OuterVolumeSpecName: "kube-api-access-k9h5w") pod "08dca185-ba7a-439d-96ee-2e5aba753dc5" (UID: "08dca185-ba7a-439d-96ee-2e5aba753dc5"). InnerVolumeSpecName "kube-api-access-k9h5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:52 crc kubenswrapper[4799]: I0930 14:41:52.005924 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9h5w\" (UniqueName: \"kubernetes.io/projected/08dca185-ba7a-439d-96ee-2e5aba753dc5-kube-api-access-k9h5w\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:52 crc kubenswrapper[4799]: I0930 14:41:52.455536 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2670-account-create-z7z29" event={"ID":"08dca185-ba7a-439d-96ee-2e5aba753dc5","Type":"ContainerDied","Data":"79cd3d4d913c03471789adb2ca8bc6de40ce99da307de3e8c0895e89fd93e400"} Sep 30 14:41:52 crc kubenswrapper[4799]: I0930 14:41:52.455625 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79cd3d4d913c03471789adb2ca8bc6de40ce99da307de3e8c0895e89fd93e400" Sep 30 14:41:52 crc kubenswrapper[4799]: I0930 14:41:52.455729 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2670-account-create-z7z29" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.473993 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerStarted","Data":"e400d2c05e69e3589bef5450ad1addec29192ec4857b5fd384600d77599bbe3d"} Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.474366 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.474210 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="ceilometer-central-agent" containerID="cri-o://32fa7121b9a6200b71434ee14f58f183fde97f26fa021511cdaf7794213b491d" gracePeriod=30 Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.474211 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="proxy-httpd" containerID="cri-o://e400d2c05e69e3589bef5450ad1addec29192ec4857b5fd384600d77599bbe3d" gracePeriod=30 Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.474288 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="sg-core" containerID="cri-o://713617e71c53f603a2a075b33b4a2d4f713d8b1d2377b5c739ca098e76d3c79f" gracePeriod=30 Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.474270 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="ceilometer-notification-agent" containerID="cri-o://fa48171e870ac0e6c827755093aa34ec1156ebefb32dad88d1cafdc5f78e10a0" gracePeriod=30 Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.514350 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.3006904280000002 podStartE2EDuration="6.514319089s" podCreationTimestamp="2025-09-30 14:41:47 +0000 UTC" firstStartedPulling="2025-09-30 14:41:48.347949236 +0000 UTC m=+1330.431549663" lastFinishedPulling="2025-09-30 14:41:52.561577897 +0000 UTC m=+1334.645178324" observedRunningTime="2025-09-30 14:41:53.506797779 +0000 UTC m=+1335.590398216" watchObservedRunningTime="2025-09-30 14:41:53.514319089 +0000 UTC m=+1335.597919516" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.897432 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rtflg"] Sep 30 14:41:53 crc kubenswrapper[4799]: E0930 14:41:53.898458 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08dca185-ba7a-439d-96ee-2e5aba753dc5" containerName="mariadb-account-create" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.898486 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="08dca185-ba7a-439d-96ee-2e5aba753dc5" containerName="mariadb-account-create" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.898738 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="08dca185-ba7a-439d-96ee-2e5aba753dc5" containerName="mariadb-account-create" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.899551 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.902846 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.904690 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.906842 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-b4n27" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.934763 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rtflg"] Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.959401 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.959480 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-scripts\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.959590 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-config-data\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:53 crc kubenswrapper[4799]: I0930 14:41:53.959620 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r77p9\" (UniqueName: \"kubernetes.io/projected/5e2fe6ff-b58e-4554-8ed1-878001b33af1-kube-api-access-r77p9\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.063325 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-config-data\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.063381 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r77p9\" (UniqueName: \"kubernetes.io/projected/5e2fe6ff-b58e-4554-8ed1-878001b33af1-kube-api-access-r77p9\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.063556 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.063593 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-scripts\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.074993 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-config-data\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.081753 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-scripts\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.088440 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.097260 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r77p9\" (UniqueName: \"kubernetes.io/projected/5e2fe6ff-b58e-4554-8ed1-878001b33af1-kube-api-access-r77p9\") pod \"nova-cell0-conductor-db-sync-rtflg\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.220807 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.550417 4799 generic.go:334] "Generic (PLEG): container finished" podID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerID="e400d2c05e69e3589bef5450ad1addec29192ec4857b5fd384600d77599bbe3d" exitCode=0 Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.550765 4799 generic.go:334] "Generic (PLEG): container finished" podID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerID="713617e71c53f603a2a075b33b4a2d4f713d8b1d2377b5c739ca098e76d3c79f" exitCode=2 Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.550779 4799 generic.go:334] "Generic (PLEG): container finished" podID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerID="32fa7121b9a6200b71434ee14f58f183fde97f26fa021511cdaf7794213b491d" exitCode=0 Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.646149 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerDied","Data":"e400d2c05e69e3589bef5450ad1addec29192ec4857b5fd384600d77599bbe3d"} Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.646549 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerDied","Data":"713617e71c53f603a2a075b33b4a2d4f713d8b1d2377b5c739ca098e76d3c79f"} Sep 30 14:41:54 crc kubenswrapper[4799]: I0930 14:41:54.646565 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerDied","Data":"32fa7121b9a6200b71434ee14f58f183fde97f26fa021511cdaf7794213b491d"} Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.051748 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rtflg"] Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.577919 4799 generic.go:334] "Generic (PLEG): container finished" podID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerID="fa48171e870ac0e6c827755093aa34ec1156ebefb32dad88d1cafdc5f78e10a0" exitCode=0 Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.577974 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerDied","Data":"fa48171e870ac0e6c827755093aa34ec1156ebefb32dad88d1cafdc5f78e10a0"} Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.578512 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ba1bdf4c-e42d-4da6-af94-9c6f23569426","Type":"ContainerDied","Data":"5fa943f02c2c00583aff101e622149e944220948f2297a4e76f8d9443bb163ff"} Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.578569 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fa943f02c2c00583aff101e622149e944220948f2297a4e76f8d9443bb163ff" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.580113 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-rtflg" event={"ID":"5e2fe6ff-b58e-4554-8ed1-878001b33af1","Type":"ContainerStarted","Data":"83527d2dda6089ebb17f64e8aa354eb0c141f3aa878ddd5f556abf2bfbafc719"} Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.620205 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.722892 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f64j\" (UniqueName: \"kubernetes.io/projected/ba1bdf4c-e42d-4da6-af94-9c6f23569426-kube-api-access-2f64j\") pod \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.723020 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-combined-ca-bundle\") pod \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.723108 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-scripts\") pod \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.727556 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-run-httpd\") pod \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.727770 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-sg-core-conf-yaml\") pod \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.727873 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-config-data\") pod \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.727935 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-log-httpd\") pod \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\" (UID: \"ba1bdf4c-e42d-4da6-af94-9c6f23569426\") " Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.729047 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ba1bdf4c-e42d-4da6-af94-9c6f23569426" (UID: "ba1bdf4c-e42d-4da6-af94-9c6f23569426"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.729772 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ba1bdf4c-e42d-4da6-af94-9c6f23569426" (UID: "ba1bdf4c-e42d-4da6-af94-9c6f23569426"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.743939 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba1bdf4c-e42d-4da6-af94-9c6f23569426-kube-api-access-2f64j" (OuterVolumeSpecName: "kube-api-access-2f64j") pod "ba1bdf4c-e42d-4da6-af94-9c6f23569426" (UID: "ba1bdf4c-e42d-4da6-af94-9c6f23569426"). InnerVolumeSpecName "kube-api-access-2f64j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.744170 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-scripts" (OuterVolumeSpecName: "scripts") pod "ba1bdf4c-e42d-4da6-af94-9c6f23569426" (UID: "ba1bdf4c-e42d-4da6-af94-9c6f23569426"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.783739 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ba1bdf4c-e42d-4da6-af94-9c6f23569426" (UID: "ba1bdf4c-e42d-4da6-af94-9c6f23569426"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.831522 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.831573 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f64j\" (UniqueName: \"kubernetes.io/projected/ba1bdf4c-e42d-4da6-af94-9c6f23569426-kube-api-access-2f64j\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.831592 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.831603 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ba1bdf4c-e42d-4da6-af94-9c6f23569426-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.831614 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.873862 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba1bdf4c-e42d-4da6-af94-9c6f23569426" (UID: "ba1bdf4c-e42d-4da6-af94-9c6f23569426"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.912155 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-config-data" (OuterVolumeSpecName: "config-data") pod "ba1bdf4c-e42d-4da6-af94-9c6f23569426" (UID: "ba1bdf4c-e42d-4da6-af94-9c6f23569426"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.933110 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:55 crc kubenswrapper[4799]: I0930 14:41:55.933154 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1bdf4c-e42d-4da6-af94-9c6f23569426-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.591033 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.623048 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.640549 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.658339 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:56 crc kubenswrapper[4799]: E0930 14:41:56.658845 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="ceilometer-notification-agent" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.658869 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="ceilometer-notification-agent" Sep 30 14:41:56 crc kubenswrapper[4799]: E0930 14:41:56.658904 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="proxy-httpd" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.658912 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="proxy-httpd" Sep 30 14:41:56 crc kubenswrapper[4799]: E0930 14:41:56.658950 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="ceilometer-central-agent" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.658960 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="ceilometer-central-agent" Sep 30 14:41:56 crc kubenswrapper[4799]: E0930 14:41:56.658974 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="sg-core" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.658982 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="sg-core" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.659303 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="ceilometer-notification-agent" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.659321 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="ceilometer-central-agent" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.659338 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="sg-core" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.659357 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" containerName="proxy-httpd" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.662963 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.667119 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.668082 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.668757 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.751537 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.751636 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-log-httpd\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.751679 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-scripts\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.751736 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jzqh\" (UniqueName: \"kubernetes.io/projected/a0553e25-b483-4ced-898c-6c466cddbdb2-kube-api-access-2jzqh\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.751757 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-run-httpd\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.751816 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.751887 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-config-data\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-log-httpd\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-scripts\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854221 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jzqh\" (UniqueName: \"kubernetes.io/projected/a0553e25-b483-4ced-898c-6c466cddbdb2-kube-api-access-2jzqh\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854250 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-run-httpd\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854315 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854386 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-config-data\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854443 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854618 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-log-httpd\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.854950 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-run-httpd\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.866140 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.866278 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-config-data\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.871919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-scripts\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.874109 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jzqh\" (UniqueName: \"kubernetes.io/projected/a0553e25-b483-4ced-898c-6c466cddbdb2-kube-api-access-2jzqh\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.878197 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " pod="openstack/ceilometer-0" Sep 30 14:41:56 crc kubenswrapper[4799]: I0930 14:41:56.990846 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:41:57 crc kubenswrapper[4799]: I0930 14:41:57.529635 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:41:57 crc kubenswrapper[4799]: I0930 14:41:57.530565 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:41:57 crc kubenswrapper[4799]: I0930 14:41:57.531853 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"2e60bf33011afb729f10179f6e079df29dfb26c28596d7ab55738fe40269d6c7"} pod="openstack/horizon-7f99689854-glm7t" containerMessage="Container horizon failed startup probe, will be restarted" Sep 30 14:41:57 crc kubenswrapper[4799]: I0930 14:41:57.531949 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" containerID="cri-o://2e60bf33011afb729f10179f6e079df29dfb26c28596d7ab55738fe40269d6c7" gracePeriod=30 Sep 30 14:41:57 crc kubenswrapper[4799]: I0930 14:41:57.557416 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:41:57 crc kubenswrapper[4799]: I0930 14:41:57.626733 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerStarted","Data":"483381d5c9a9fe6a79b0bb50834fbdd29e14245b0648b401067cb0d7fd8034b7"} Sep 30 14:41:58 crc kubenswrapper[4799]: I0930 14:41:58.531026 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba1bdf4c-e42d-4da6-af94-9c6f23569426" path="/var/lib/kubelet/pods/ba1bdf4c-e42d-4da6-af94-9c6f23569426/volumes" Sep 30 14:41:59 crc kubenswrapper[4799]: I0930 14:41:59.649053 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:41:59 crc kubenswrapper[4799]: I0930 14:41:59.649421 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:41:59 crc kubenswrapper[4799]: I0930 14:41:59.649539 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:41:59 crc kubenswrapper[4799]: I0930 14:41:59.650334 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb9ad6936dc7236258e4efd5f74222bbe45a21ab72fed8d7b84c8d34eafe641b"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:41:59 crc kubenswrapper[4799]: I0930 14:41:59.650387 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://bb9ad6936dc7236258e4efd5f74222bbe45a21ab72fed8d7b84c8d34eafe641b" gracePeriod=600 Sep 30 14:41:59 crc kubenswrapper[4799]: I0930 14:41:59.662991 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerStarted","Data":"386c69a6d45ba00c080c113e1e249f3079ecb29291046b293f941e409c9e24b3"} Sep 30 14:42:00 crc kubenswrapper[4799]: I0930 14:42:00.679838 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="bb9ad6936dc7236258e4efd5f74222bbe45a21ab72fed8d7b84c8d34eafe641b" exitCode=0 Sep 30 14:42:00 crc kubenswrapper[4799]: I0930 14:42:00.680399 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"bb9ad6936dc7236258e4efd5f74222bbe45a21ab72fed8d7b84c8d34eafe641b"} Sep 30 14:42:00 crc kubenswrapper[4799]: I0930 14:42:00.680454 4799 scope.go:117] "RemoveContainer" containerID="c08cc75dc1edb21f93545a035f6f554f287534552d52a3400113832177d88167" Sep 30 14:42:01 crc kubenswrapper[4799]: I0930 14:42:01.723000 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:42:01 crc kubenswrapper[4799]: I0930 14:42:01.723743 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerName="glance-log" containerID="cri-o://6fe6ceb0e8172a311716a0c0048b38d77e10ec63ad8a357eabc21d83f5f7e320" gracePeriod=30 Sep 30 14:42:01 crc kubenswrapper[4799]: I0930 14:42:01.724398 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerName="glance-httpd" containerID="cri-o://6cb6f4203908431a077c358319044c348ecce50af75f2791408d4e0f743a8a8a" gracePeriod=30 Sep 30 14:42:02 crc kubenswrapper[4799]: I0930 14:42:02.716302 4799 generic.go:334] "Generic (PLEG): container finished" podID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerID="6fe6ceb0e8172a311716a0c0048b38d77e10ec63ad8a357eabc21d83f5f7e320" exitCode=143 Sep 30 14:42:02 crc kubenswrapper[4799]: I0930 14:42:02.716370 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"263ea967-0679-4f9f-81f3-4f1eb6b20835","Type":"ContainerDied","Data":"6fe6ceb0e8172a311716a0c0048b38d77e10ec63ad8a357eabc21d83f5f7e320"} Sep 30 14:42:02 crc kubenswrapper[4799]: I0930 14:42:02.805079 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:42:02 crc kubenswrapper[4799]: I0930 14:42:02.805618 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="27e3a51e-4278-4183-8329-15916d7304a7" containerName="glance-log" containerID="cri-o://ee708f9e443e77270b2fa71111914e5caec7831065bb86be29b62ea5f87db5e6" gracePeriod=30 Sep 30 14:42:02 crc kubenswrapper[4799]: I0930 14:42:02.812471 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="27e3a51e-4278-4183-8329-15916d7304a7" containerName="glance-httpd" containerID="cri-o://d58da57a64dea1a04e5a5194dace7d97a3343dd15307cb9b3a2ca9c255f06efa" gracePeriod=30 Sep 30 14:42:03 crc kubenswrapper[4799]: I0930 14:42:03.739153 4799 generic.go:334] "Generic (PLEG): container finished" podID="27e3a51e-4278-4183-8329-15916d7304a7" containerID="ee708f9e443e77270b2fa71111914e5caec7831065bb86be29b62ea5f87db5e6" exitCode=143 Sep 30 14:42:03 crc kubenswrapper[4799]: I0930 14:42:03.739354 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"27e3a51e-4278-4183-8329-15916d7304a7","Type":"ContainerDied","Data":"ee708f9e443e77270b2fa71111914e5caec7831065bb86be29b62ea5f87db5e6"} Sep 30 14:42:04 crc kubenswrapper[4799]: I0930 14:42:04.046542 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:42:05 crc kubenswrapper[4799]: I0930 14:42:05.786534 4799 generic.go:334] "Generic (PLEG): container finished" podID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerID="6cb6f4203908431a077c358319044c348ecce50af75f2791408d4e0f743a8a8a" exitCode=0 Sep 30 14:42:05 crc kubenswrapper[4799]: I0930 14:42:05.787158 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"263ea967-0679-4f9f-81f3-4f1eb6b20835","Type":"ContainerDied","Data":"6cb6f4203908431a077c358319044c348ecce50af75f2791408d4e0f743a8a8a"} Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.072495 4799 generic.go:334] "Generic (PLEG): container finished" podID="27e3a51e-4278-4183-8329-15916d7304a7" containerID="d58da57a64dea1a04e5a5194dace7d97a3343dd15307cb9b3a2ca9c255f06efa" exitCode=0 Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.073020 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"27e3a51e-4278-4183-8329-15916d7304a7","Type":"ContainerDied","Data":"d58da57a64dea1a04e5a5194dace7d97a3343dd15307cb9b3a2ca9c255f06efa"} Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.251788 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.287356 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-scripts\") pod \"263ea967-0679-4f9f-81f3-4f1eb6b20835\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.287454 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-combined-ca-bundle\") pod \"263ea967-0679-4f9f-81f3-4f1eb6b20835\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.287518 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-httpd-run\") pod \"263ea967-0679-4f9f-81f3-4f1eb6b20835\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.287558 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-config-data\") pod \"263ea967-0679-4f9f-81f3-4f1eb6b20835\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.287587 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"263ea967-0679-4f9f-81f3-4f1eb6b20835\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.287632 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c542m\" (UniqueName: \"kubernetes.io/projected/263ea967-0679-4f9f-81f3-4f1eb6b20835-kube-api-access-c542m\") pod \"263ea967-0679-4f9f-81f3-4f1eb6b20835\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.287707 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-logs\") pod \"263ea967-0679-4f9f-81f3-4f1eb6b20835\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.287759 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-public-tls-certs\") pod \"263ea967-0679-4f9f-81f3-4f1eb6b20835\" (UID: \"263ea967-0679-4f9f-81f3-4f1eb6b20835\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.301097 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-logs" (OuterVolumeSpecName: "logs") pod "263ea967-0679-4f9f-81f3-4f1eb6b20835" (UID: "263ea967-0679-4f9f-81f3-4f1eb6b20835"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.314289 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "263ea967-0679-4f9f-81f3-4f1eb6b20835" (UID: "263ea967-0679-4f9f-81f3-4f1eb6b20835"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.314536 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "263ea967-0679-4f9f-81f3-4f1eb6b20835" (UID: "263ea967-0679-4f9f-81f3-4f1eb6b20835"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.325407 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-scripts" (OuterVolumeSpecName: "scripts") pod "263ea967-0679-4f9f-81f3-4f1eb6b20835" (UID: "263ea967-0679-4f9f-81f3-4f1eb6b20835"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.333334 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/263ea967-0679-4f9f-81f3-4f1eb6b20835-kube-api-access-c542m" (OuterVolumeSpecName: "kube-api-access-c542m") pod "263ea967-0679-4f9f-81f3-4f1eb6b20835" (UID: "263ea967-0679-4f9f-81f3-4f1eb6b20835"). InnerVolumeSpecName "kube-api-access-c542m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.360020 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "263ea967-0679-4f9f-81f3-4f1eb6b20835" (UID: "263ea967-0679-4f9f-81f3-4f1eb6b20835"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.398281 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c542m\" (UniqueName: \"kubernetes.io/projected/263ea967-0679-4f9f-81f3-4f1eb6b20835-kube-api-access-c542m\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.398321 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.398334 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.398345 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.398355 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/263ea967-0679-4f9f-81f3-4f1eb6b20835-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.402749 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.452733 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.511247 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.595917 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "263ea967-0679-4f9f-81f3-4f1eb6b20835" (UID: "263ea967-0679-4f9f-81f3-4f1eb6b20835"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.630435 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.634808 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.651921 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-config-data" (OuterVolumeSpecName: "config-data") pod "263ea967-0679-4f9f-81f3-4f1eb6b20835" (UID: "263ea967-0679-4f9f-81f3-4f1eb6b20835"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.731525 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-logs\") pod \"27e3a51e-4278-4183-8329-15916d7304a7\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.731990 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-scripts\") pod \"27e3a51e-4278-4183-8329-15916d7304a7\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.732593 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-logs" (OuterVolumeSpecName: "logs") pod "27e3a51e-4278-4183-8329-15916d7304a7" (UID: "27e3a51e-4278-4183-8329-15916d7304a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.733121 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-httpd-run\") pod \"27e3a51e-4278-4183-8329-15916d7304a7\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.733296 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-internal-tls-certs\") pod \"27e3a51e-4278-4183-8329-15916d7304a7\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.733404 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"27e3a51e-4278-4183-8329-15916d7304a7\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.733536 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnj2r\" (UniqueName: \"kubernetes.io/projected/27e3a51e-4278-4183-8329-15916d7304a7-kube-api-access-vnj2r\") pod \"27e3a51e-4278-4183-8329-15916d7304a7\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.733332 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "27e3a51e-4278-4183-8329-15916d7304a7" (UID: "27e3a51e-4278-4183-8329-15916d7304a7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.733704 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-config-data\") pod \"27e3a51e-4278-4183-8329-15916d7304a7\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.733872 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-combined-ca-bundle\") pod \"27e3a51e-4278-4183-8329-15916d7304a7\" (UID: \"27e3a51e-4278-4183-8329-15916d7304a7\") " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.734572 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.734599 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/263ea967-0679-4f9f-81f3-4f1eb6b20835-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.734613 4799 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27e3a51e-4278-4183-8329-15916d7304a7-httpd-run\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.745573 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "27e3a51e-4278-4183-8329-15916d7304a7" (UID: "27e3a51e-4278-4183-8329-15916d7304a7"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.775361 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-scripts" (OuterVolumeSpecName: "scripts") pod "27e3a51e-4278-4183-8329-15916d7304a7" (UID: "27e3a51e-4278-4183-8329-15916d7304a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.783221 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e3a51e-4278-4183-8329-15916d7304a7-kube-api-access-vnj2r" (OuterVolumeSpecName: "kube-api-access-vnj2r") pod "27e3a51e-4278-4183-8329-15916d7304a7" (UID: "27e3a51e-4278-4183-8329-15916d7304a7"). InnerVolumeSpecName "kube-api-access-vnj2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.809121 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27e3a51e-4278-4183-8329-15916d7304a7" (UID: "27e3a51e-4278-4183-8329-15916d7304a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.852037 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.852080 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.852122 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.852136 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnj2r\" (UniqueName: \"kubernetes.io/projected/27e3a51e-4278-4183-8329-15916d7304a7-kube-api-access-vnj2r\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.917405 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.926537 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "27e3a51e-4278-4183-8329-15916d7304a7" (UID: "27e3a51e-4278-4183-8329-15916d7304a7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.957357 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.957749 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:07 crc kubenswrapper[4799]: I0930 14:42:07.973808 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-config-data" (OuterVolumeSpecName: "config-data") pod "27e3a51e-4278-4183-8329-15916d7304a7" (UID: "27e3a51e-4278-4183-8329-15916d7304a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.060117 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27e3a51e-4278-4183-8329-15916d7304a7-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.106630 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"27e3a51e-4278-4183-8329-15916d7304a7","Type":"ContainerDied","Data":"be3e4333524dfa827545ef0ce8b5f69d33fcf536316bfa9742f9e4aaf44edc8d"} Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.107806 4799 scope.go:117] "RemoveContainer" containerID="d58da57a64dea1a04e5a5194dace7d97a3343dd15307cb9b3a2ca9c255f06efa" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.107852 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.117799 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"263ea967-0679-4f9f-81f3-4f1eb6b20835","Type":"ContainerDied","Data":"0ae4407a355479943f2fa9625694789170d4eeb69fc4ae873f199fab6152f230"} Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.117919 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.131577 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerStarted","Data":"090b59469bf84d3f8c9547066c5a9c9b3b624194687de7d9b23629a9d56c8aeb"} Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.151700 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-rtflg" event={"ID":"5e2fe6ff-b58e-4554-8ed1-878001b33af1","Type":"ContainerStarted","Data":"7164927e50e022176906560d22458e8ff4410f75da5bcbc2fd907d362fb9d9a1"} Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.182971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1"} Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.183071 4799 scope.go:117] "RemoveContainer" containerID="ee708f9e443e77270b2fa71111914e5caec7831065bb86be29b62ea5f87db5e6" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.257624 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-rtflg" podStartSLOduration=3.611480931 podStartE2EDuration="15.257595992s" podCreationTimestamp="2025-09-30 14:41:53 +0000 UTC" firstStartedPulling="2025-09-30 14:41:55.09162103 +0000 UTC m=+1337.175221457" lastFinishedPulling="2025-09-30 14:42:06.737736081 +0000 UTC m=+1348.821336518" observedRunningTime="2025-09-30 14:42:08.250260848 +0000 UTC m=+1350.333861305" watchObservedRunningTime="2025-09-30 14:42:08.257595992 +0000 UTC m=+1350.341196419" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.268899 4799 scope.go:117] "RemoveContainer" containerID="6cb6f4203908431a077c358319044c348ecce50af75f2791408d4e0f743a8a8a" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.292822 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.341376 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.350127 4799 scope.go:117] "RemoveContainer" containerID="6fe6ceb0e8172a311716a0c0048b38d77e10ec63ad8a357eabc21d83f5f7e320" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.353801 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.379777 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.405514 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:42:08 crc kubenswrapper[4799]: E0930 14:42:08.406129 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerName="glance-httpd" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.406161 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerName="glance-httpd" Sep 30 14:42:08 crc kubenswrapper[4799]: E0930 14:42:08.406185 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e3a51e-4278-4183-8329-15916d7304a7" containerName="glance-httpd" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.406193 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e3a51e-4278-4183-8329-15916d7304a7" containerName="glance-httpd" Sep 30 14:42:08 crc kubenswrapper[4799]: E0930 14:42:08.406221 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e3a51e-4278-4183-8329-15916d7304a7" containerName="glance-log" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.406230 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e3a51e-4278-4183-8329-15916d7304a7" containerName="glance-log" Sep 30 14:42:08 crc kubenswrapper[4799]: E0930 14:42:08.406252 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerName="glance-log" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.406260 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerName="glance-log" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.406497 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerName="glance-httpd" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.406520 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" containerName="glance-log" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.406548 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e3a51e-4278-4183-8329-15916d7304a7" containerName="glance-log" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.406561 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e3a51e-4278-4183-8329-15916d7304a7" containerName="glance-httpd" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.408034 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.416422 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.416771 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.416941 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pjthr" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.417337 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.439279 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.441383 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.447236 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.482988 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.483424 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.530602 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="263ea967-0679-4f9f-81f3-4f1eb6b20835" path="/var/lib/kubelet/pods/263ea967-0679-4f9f-81f3-4f1eb6b20835/volumes" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.536622 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27e3a51e-4278-4183-8329-15916d7304a7" path="/var/lib/kubelet/pods/27e3a51e-4278-4183-8329-15916d7304a7/volumes" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.537680 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.574300 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bbb01bf-9925-4813-94e7-17882aa4d4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.574373 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.574467 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.574508 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bbb01bf-9925-4813-94e7-17882aa4d4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.574590 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.574717 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.574756 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9w4n\" (UniqueName: \"kubernetes.io/projected/5bbb01bf-9925-4813-94e7-17882aa4d4c5-kube-api-access-g9w4n\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.574798 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.676962 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-scripts\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677064 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677096 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-config-data\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677164 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bbb01bf-9925-4813-94e7-17882aa4d4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677278 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677338 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677383 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9w4n\" (UniqueName: \"kubernetes.io/projected/5bbb01bf-9925-4813-94e7-17882aa4d4c5-kube-api-access-g9w4n\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677424 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0334f39c-a7d6-42f3-a764-04e8e4857c41-logs\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677496 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677569 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677689 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bbb01bf-9925-4813-94e7-17882aa4d4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677728 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677832 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0334f39c-a7d6-42f3-a764-04e8e4857c41-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677862 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.677908 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jkrv\" (UniqueName: \"kubernetes.io/projected/0334f39c-a7d6-42f3-a764-04e8e4857c41-kube-api-access-2jkrv\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.678860 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.681626 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5bbb01bf-9925-4813-94e7-17882aa4d4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.686712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bbb01bf-9925-4813-94e7-17882aa4d4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.705093 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.712319 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.715463 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.716423 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bbb01bf-9925-4813-94e7-17882aa4d4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.794719 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.794799 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jkrv\" (UniqueName: \"kubernetes.io/projected/0334f39c-a7d6-42f3-a764-04e8e4857c41-kube-api-access-2jkrv\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.794880 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-scripts\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.794910 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-config-data\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.795046 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0334f39c-a7d6-42f3-a764-04e8e4857c41-logs\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.795109 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.795135 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.795245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0334f39c-a7d6-42f3-a764-04e8e4857c41-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.805246 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.812434 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.812901 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0334f39c-a7d6-42f3-a764-04e8e4857c41-logs\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.814385 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0334f39c-a7d6-42f3-a764-04e8e4857c41-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.855698 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-scripts\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.861207 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.862596 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0334f39c-a7d6-42f3-a764-04e8e4857c41-config-data\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.905317 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jkrv\" (UniqueName: \"kubernetes.io/projected/0334f39c-a7d6-42f3-a764-04e8e4857c41-kube-api-access-2jkrv\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.916506 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9w4n\" (UniqueName: \"kubernetes.io/projected/5bbb01bf-9925-4813-94e7-17882aa4d4c5-kube-api-access-g9w4n\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:08 crc kubenswrapper[4799]: I0930 14:42:08.953577 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"5bbb01bf-9925-4813-94e7-17882aa4d4c5\") " pod="openstack/glance-default-internal-api-0" Sep 30 14:42:09 crc kubenswrapper[4799]: I0930 14:42:09.005239 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0334f39c-a7d6-42f3-a764-04e8e4857c41\") " pod="openstack/glance-default-external-api-0" Sep 30 14:42:09 crc kubenswrapper[4799]: I0930 14:42:09.040604 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:09 crc kubenswrapper[4799]: I0930 14:42:09.079187 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Sep 30 14:42:09 crc kubenswrapper[4799]: I0930 14:42:09.985043 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Sep 30 14:42:10 crc kubenswrapper[4799]: W0930 14:42:10.009198 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bbb01bf_9925_4813_94e7_17882aa4d4c5.slice/crio-3bebcc671ceae1a08ffcc1e87264371d9986db0ad73f15946fdaaa87147c6c25 WatchSource:0}: Error finding container 3bebcc671ceae1a08ffcc1e87264371d9986db0ad73f15946fdaaa87147c6c25: Status 404 returned error can't find the container with id 3bebcc671ceae1a08ffcc1e87264371d9986db0ad73f15946fdaaa87147c6c25 Sep 30 14:42:10 crc kubenswrapper[4799]: I0930 14:42:10.268363 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerStarted","Data":"584dbdd8e0cda1ee7f7710b4e068a19d2fc473344da7e9604a9166774c43ed45"} Sep 30 14:42:10 crc kubenswrapper[4799]: I0930 14:42:10.274784 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bbb01bf-9925-4813-94e7-17882aa4d4c5","Type":"ContainerStarted","Data":"3bebcc671ceae1a08ffcc1e87264371d9986db0ad73f15946fdaaa87147c6c25"} Sep 30 14:42:10 crc kubenswrapper[4799]: I0930 14:42:10.327126 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Sep 30 14:42:11 crc kubenswrapper[4799]: I0930 14:42:11.301408 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0334f39c-a7d6-42f3-a764-04e8e4857c41","Type":"ContainerStarted","Data":"38e11bcac50d04d50ab6ed55441313375efa544090f1d8f54a4d9ac2c47ab0c1"} Sep 30 14:42:11 crc kubenswrapper[4799]: I0930 14:42:11.323684 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerStarted","Data":"654562b9be47c9346b3b506e1f6f79cbf36a6eb2fc613844bdb5f101bebb85e7"} Sep 30 14:42:11 crc kubenswrapper[4799]: I0930 14:42:11.323913 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="ceilometer-central-agent" containerID="cri-o://386c69a6d45ba00c080c113e1e249f3079ecb29291046b293f941e409c9e24b3" gracePeriod=30 Sep 30 14:42:11 crc kubenswrapper[4799]: I0930 14:42:11.324028 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:42:11 crc kubenswrapper[4799]: I0930 14:42:11.324432 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="proxy-httpd" containerID="cri-o://654562b9be47c9346b3b506e1f6f79cbf36a6eb2fc613844bdb5f101bebb85e7" gracePeriod=30 Sep 30 14:42:11 crc kubenswrapper[4799]: I0930 14:42:11.324483 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="sg-core" containerID="cri-o://584dbdd8e0cda1ee7f7710b4e068a19d2fc473344da7e9604a9166774c43ed45" gracePeriod=30 Sep 30 14:42:11 crc kubenswrapper[4799]: I0930 14:42:11.324532 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="ceilometer-notification-agent" containerID="cri-o://090b59469bf84d3f8c9547066c5a9c9b3b624194687de7d9b23629a9d56c8aeb" gracePeriod=30 Sep 30 14:42:11 crc kubenswrapper[4799]: I0930 14:42:11.383206 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.425559733 podStartE2EDuration="15.383176816s" podCreationTimestamp="2025-09-30 14:41:56 +0000 UTC" firstStartedPulling="2025-09-30 14:41:57.590332317 +0000 UTC m=+1339.673932744" lastFinishedPulling="2025-09-30 14:42:10.5479494 +0000 UTC m=+1352.631549827" observedRunningTime="2025-09-30 14:42:11.358501915 +0000 UTC m=+1353.442102342" watchObservedRunningTime="2025-09-30 14:42:11.383176816 +0000 UTC m=+1353.466777243" Sep 30 14:42:12 crc kubenswrapper[4799]: I0930 14:42:12.358389 4799 generic.go:334] "Generic (PLEG): container finished" podID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerID="584dbdd8e0cda1ee7f7710b4e068a19d2fc473344da7e9604a9166774c43ed45" exitCode=2 Sep 30 14:42:12 crc kubenswrapper[4799]: I0930 14:42:12.358797 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerDied","Data":"584dbdd8e0cda1ee7f7710b4e068a19d2fc473344da7e9604a9166774c43ed45"} Sep 30 14:42:12 crc kubenswrapper[4799]: I0930 14:42:12.369627 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0334f39c-a7d6-42f3-a764-04e8e4857c41","Type":"ContainerStarted","Data":"03efefc753243b35fa8aad4965d30fdf098a669966fbb123ac6d56cfd2596156"} Sep 30 14:42:12 crc kubenswrapper[4799]: I0930 14:42:12.376263 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bbb01bf-9925-4813-94e7-17882aa4d4c5","Type":"ContainerStarted","Data":"c67e41ea916fc34cdb77e178ffa5e8b8139327a53a60f74e4eb8fe68e0a5a667"} Sep 30 14:42:13 crc kubenswrapper[4799]: I0930 14:42:13.396927 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0334f39c-a7d6-42f3-a764-04e8e4857c41","Type":"ContainerStarted","Data":"30681e60d1244bd8bcc18a9a503ca25cb345c2b2b266ebdea4638cf4413429bd"} Sep 30 14:42:13 crc kubenswrapper[4799]: I0930 14:42:13.399351 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5bbb01bf-9925-4813-94e7-17882aa4d4c5","Type":"ContainerStarted","Data":"aec9bbf9526fb4e5921ae151c533959a08a3d1caadb82bc6d766ae3d5af33f3d"} Sep 30 14:42:13 crc kubenswrapper[4799]: I0930 14:42:13.404271 4799 generic.go:334] "Generic (PLEG): container finished" podID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerID="386c69a6d45ba00c080c113e1e249f3079ecb29291046b293f941e409c9e24b3" exitCode=0 Sep 30 14:42:13 crc kubenswrapper[4799]: I0930 14:42:13.404329 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerDied","Data":"386c69a6d45ba00c080c113e1e249f3079ecb29291046b293f941e409c9e24b3"} Sep 30 14:42:14 crc kubenswrapper[4799]: I0930 14:42:14.454345 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.454310228 podStartE2EDuration="6.454310228s" podCreationTimestamp="2025-09-30 14:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:42:14.451384513 +0000 UTC m=+1356.534984960" watchObservedRunningTime="2025-09-30 14:42:14.454310228 +0000 UTC m=+1356.537910655" Sep 30 14:42:14 crc kubenswrapper[4799]: I0930 14:42:14.502227 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.502196858 podStartE2EDuration="6.502196858s" podCreationTimestamp="2025-09-30 14:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:42:14.489317792 +0000 UTC m=+1356.572918219" watchObservedRunningTime="2025-09-30 14:42:14.502196858 +0000 UTC m=+1356.585797275" Sep 30 14:42:18 crc kubenswrapper[4799]: I0930 14:42:18.486234 4799 generic.go:334] "Generic (PLEG): container finished" podID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerID="2aa2de6ee9c44af7d31b1105ea4177bb893946376162ccd5d747ef4d0d8305e3" exitCode=137 Sep 30 14:42:18 crc kubenswrapper[4799]: I0930 14:42:18.486312 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerDied","Data":"2aa2de6ee9c44af7d31b1105ea4177bb893946376162ccd5d747ef4d0d8305e3"} Sep 30 14:42:18 crc kubenswrapper[4799]: I0930 14:42:18.486960 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerStarted","Data":"145eff7bbb6dd0a52fa1de14ab2ef3666affe0c5dae478a2644a1c0dffeb1113"} Sep 30 14:42:18 crc kubenswrapper[4799]: I0930 14:42:18.486992 4799 scope.go:117] "RemoveContainer" containerID="0db7b3053e7c7468a8690c8b47ee3b620779c70dce53efef8b358e92bdd7e983" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.041937 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.042005 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.079737 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.080316 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.106492 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.116861 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.131794 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.157770 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.510780 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.510870 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.511368 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:19 crc kubenswrapper[4799]: I0930 14:42:19.511785 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:21 crc kubenswrapper[4799]: I0930 14:42:21.534259 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:42:21 crc kubenswrapper[4799]: I0930 14:42:21.534319 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:42:23 crc kubenswrapper[4799]: I0930 14:42:23.763349 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 14:42:23 crc kubenswrapper[4799]: I0930 14:42:23.763941 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:42:23 crc kubenswrapper[4799]: I0930 14:42:23.767289 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Sep 30 14:42:23 crc kubenswrapper[4799]: I0930 14:42:23.780776 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:23 crc kubenswrapper[4799]: I0930 14:42:23.781531 4799 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 30 14:42:23 crc kubenswrapper[4799]: I0930 14:42:23.785536 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Sep 30 14:42:26 crc kubenswrapper[4799]: I0930 14:42:26.998496 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 30 14:42:27 crc kubenswrapper[4799]: I0930 14:42:27.681876 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:42:27 crc kubenswrapper[4799]: I0930 14:42:27.682286 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:42:28 crc kubenswrapper[4799]: I0930 14:42:28.628676 4799 generic.go:334] "Generic (PLEG): container finished" podID="d15890bb-d088-4126-8382-b5d8e423b608" containerID="2e60bf33011afb729f10179f6e079df29dfb26c28596d7ab55738fe40269d6c7" exitCode=137 Sep 30 14:42:28 crc kubenswrapper[4799]: I0930 14:42:28.628760 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerDied","Data":"2e60bf33011afb729f10179f6e079df29dfb26c28596d7ab55738fe40269d6c7"} Sep 30 14:42:28 crc kubenswrapper[4799]: I0930 14:42:28.629350 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerStarted","Data":"6a24eead33c55554d4fa163faed09d5db7e530d41a8e1d35cb03d928a56c1750"} Sep 30 14:42:28 crc kubenswrapper[4799]: I0930 14:42:28.629400 4799 scope.go:117] "RemoveContainer" containerID="6218e563ed5620135720c08ff067c9d79266769a1728d09cab51ff7af21c5dfe" Sep 30 14:42:33 crc kubenswrapper[4799]: I0930 14:42:33.707737 4799 generic.go:334] "Generic (PLEG): container finished" podID="5e2fe6ff-b58e-4554-8ed1-878001b33af1" containerID="7164927e50e022176906560d22458e8ff4410f75da5bcbc2fd907d362fb9d9a1" exitCode=0 Sep 30 14:42:33 crc kubenswrapper[4799]: I0930 14:42:33.707833 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-rtflg" event={"ID":"5e2fe6ff-b58e-4554-8ed1-878001b33af1","Type":"ContainerDied","Data":"7164927e50e022176906560d22458e8ff4410f75da5bcbc2fd907d362fb9d9a1"} Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.172624 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.241340 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-scripts\") pod \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.241642 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-config-data\") pod \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.241874 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-combined-ca-bundle\") pod \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.241995 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r77p9\" (UniqueName: \"kubernetes.io/projected/5e2fe6ff-b58e-4554-8ed1-878001b33af1-kube-api-access-r77p9\") pod \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\" (UID: \"5e2fe6ff-b58e-4554-8ed1-878001b33af1\") " Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.262027 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2fe6ff-b58e-4554-8ed1-878001b33af1-kube-api-access-r77p9" (OuterVolumeSpecName: "kube-api-access-r77p9") pod "5e2fe6ff-b58e-4554-8ed1-878001b33af1" (UID: "5e2fe6ff-b58e-4554-8ed1-878001b33af1"). InnerVolumeSpecName "kube-api-access-r77p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.297867 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-scripts" (OuterVolumeSpecName: "scripts") pod "5e2fe6ff-b58e-4554-8ed1-878001b33af1" (UID: "5e2fe6ff-b58e-4554-8ed1-878001b33af1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.316488 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e2fe6ff-b58e-4554-8ed1-878001b33af1" (UID: "5e2fe6ff-b58e-4554-8ed1-878001b33af1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.337334 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-config-data" (OuterVolumeSpecName: "config-data") pod "5e2fe6ff-b58e-4554-8ed1-878001b33af1" (UID: "5e2fe6ff-b58e-4554-8ed1-878001b33af1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.349902 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.350030 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r77p9\" (UniqueName: \"kubernetes.io/projected/5e2fe6ff-b58e-4554-8ed1-878001b33af1-kube-api-access-r77p9\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.350045 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.350055 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2fe6ff-b58e-4554-8ed1-878001b33af1-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.729198 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-rtflg" event={"ID":"5e2fe6ff-b58e-4554-8ed1-878001b33af1","Type":"ContainerDied","Data":"83527d2dda6089ebb17f64e8aa354eb0c141f3aa878ddd5f556abf2bfbafc719"} Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.729268 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83527d2dda6089ebb17f64e8aa354eb0c141f3aa878ddd5f556abf2bfbafc719" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.729329 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-rtflg" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.944240 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 14:42:35 crc kubenswrapper[4799]: E0930 14:42:35.944843 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2fe6ff-b58e-4554-8ed1-878001b33af1" containerName="nova-cell0-conductor-db-sync" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.944864 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2fe6ff-b58e-4554-8ed1-878001b33af1" containerName="nova-cell0-conductor-db-sync" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.945100 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2fe6ff-b58e-4554-8ed1-878001b33af1" containerName="nova-cell0-conductor-db-sync" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.945996 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.950418 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.960477 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-b4n27" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.965438 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.967189 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f55c7c0-30d5-4349-956a-c06acef9fba0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.967251 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7j9n\" (UniqueName: \"kubernetes.io/projected/4f55c7c0-30d5-4349-956a-c06acef9fba0-kube-api-access-q7j9n\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:35 crc kubenswrapper[4799]: I0930 14:42:35.967474 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f55c7c0-30d5-4349-956a-c06acef9fba0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:36 crc kubenswrapper[4799]: I0930 14:42:36.069491 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f55c7c0-30d5-4349-956a-c06acef9fba0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:36 crc kubenswrapper[4799]: I0930 14:42:36.069717 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f55c7c0-30d5-4349-956a-c06acef9fba0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:36 crc kubenswrapper[4799]: I0930 14:42:36.069766 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7j9n\" (UniqueName: \"kubernetes.io/projected/4f55c7c0-30d5-4349-956a-c06acef9fba0-kube-api-access-q7j9n\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:36 crc kubenswrapper[4799]: I0930 14:42:36.076807 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f55c7c0-30d5-4349-956a-c06acef9fba0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:36 crc kubenswrapper[4799]: I0930 14:42:36.079908 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f55c7c0-30d5-4349-956a-c06acef9fba0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:36 crc kubenswrapper[4799]: I0930 14:42:36.098983 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7j9n\" (UniqueName: \"kubernetes.io/projected/4f55c7c0-30d5-4349-956a-c06acef9fba0-kube-api-access-q7j9n\") pod \"nova-cell0-conductor-0\" (UID: \"4f55c7c0-30d5-4349-956a-c06acef9fba0\") " pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:36 crc kubenswrapper[4799]: I0930 14:42:36.270774 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:36 crc kubenswrapper[4799]: I0930 14:42:36.877777 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Sep 30 14:42:37 crc kubenswrapper[4799]: I0930 14:42:37.529276 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:42:37 crc kubenswrapper[4799]: I0930 14:42:37.530809 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:42:37 crc kubenswrapper[4799]: I0930 14:42:37.531272 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:42:37 crc kubenswrapper[4799]: I0930 14:42:37.684823 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:42:37 crc kubenswrapper[4799]: I0930 14:42:37.754542 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4f55c7c0-30d5-4349-956a-c06acef9fba0","Type":"ContainerStarted","Data":"e8fec0e046f0428cc81d9d97f216a8b12c61c0db672950882c1eec07d7ab26d9"} Sep 30 14:42:37 crc kubenswrapper[4799]: I0930 14:42:37.754663 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:37 crc kubenswrapper[4799]: I0930 14:42:37.754687 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4f55c7c0-30d5-4349-956a-c06acef9fba0","Type":"ContainerStarted","Data":"418a54b440b15f5388bad2921632281f02c033909104a8f5944fc30a274d3a83"} Sep 30 14:42:37 crc kubenswrapper[4799]: I0930 14:42:37.778986 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.778952318 podStartE2EDuration="2.778952318s" podCreationTimestamp="2025-09-30 14:42:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:42:37.77560816 +0000 UTC m=+1379.859208587" watchObservedRunningTime="2025-09-30 14:42:37.778952318 +0000 UTC m=+1379.862552745" Sep 30 14:42:41 crc kubenswrapper[4799]: I0930 14:42:41.815186 4799 generic.go:334] "Generic (PLEG): container finished" podID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerID="654562b9be47c9346b3b506e1f6f79cbf36a6eb2fc613844bdb5f101bebb85e7" exitCode=137 Sep 30 14:42:41 crc kubenswrapper[4799]: I0930 14:42:41.815782 4799 generic.go:334] "Generic (PLEG): container finished" podID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerID="090b59469bf84d3f8c9547066c5a9c9b3b624194687de7d9b23629a9d56c8aeb" exitCode=137 Sep 30 14:42:41 crc kubenswrapper[4799]: I0930 14:42:41.815234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerDied","Data":"654562b9be47c9346b3b506e1f6f79cbf36a6eb2fc613844bdb5f101bebb85e7"} Sep 30 14:42:41 crc kubenswrapper[4799]: I0930 14:42:41.815836 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerDied","Data":"090b59469bf84d3f8c9547066c5a9c9b3b624194687de7d9b23629a9d56c8aeb"} Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.298178 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.409392 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-sg-core-conf-yaml\") pod \"a0553e25-b483-4ced-898c-6c466cddbdb2\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.409462 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-run-httpd\") pod \"a0553e25-b483-4ced-898c-6c466cddbdb2\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.409616 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-log-httpd\") pod \"a0553e25-b483-4ced-898c-6c466cddbdb2\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.410690 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-scripts\") pod \"a0553e25-b483-4ced-898c-6c466cddbdb2\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.410702 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a0553e25-b483-4ced-898c-6c466cddbdb2" (UID: "a0553e25-b483-4ced-898c-6c466cddbdb2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.410759 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jzqh\" (UniqueName: \"kubernetes.io/projected/a0553e25-b483-4ced-898c-6c466cddbdb2-kube-api-access-2jzqh\") pod \"a0553e25-b483-4ced-898c-6c466cddbdb2\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.410952 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-combined-ca-bundle\") pod \"a0553e25-b483-4ced-898c-6c466cddbdb2\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.410985 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-config-data\") pod \"a0553e25-b483-4ced-898c-6c466cddbdb2\" (UID: \"a0553e25-b483-4ced-898c-6c466cddbdb2\") " Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.411967 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.412711 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a0553e25-b483-4ced-898c-6c466cddbdb2" (UID: "a0553e25-b483-4ced-898c-6c466cddbdb2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.426736 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-scripts" (OuterVolumeSpecName: "scripts") pod "a0553e25-b483-4ced-898c-6c466cddbdb2" (UID: "a0553e25-b483-4ced-898c-6c466cddbdb2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.438001 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0553e25-b483-4ced-898c-6c466cddbdb2-kube-api-access-2jzqh" (OuterVolumeSpecName: "kube-api-access-2jzqh") pod "a0553e25-b483-4ced-898c-6c466cddbdb2" (UID: "a0553e25-b483-4ced-898c-6c466cddbdb2"). InnerVolumeSpecName "kube-api-access-2jzqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.478810 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a0553e25-b483-4ced-898c-6c466cddbdb2" (UID: "a0553e25-b483-4ced-898c-6c466cddbdb2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.518819 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.519155 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0553e25-b483-4ced-898c-6c466cddbdb2-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.519243 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.519329 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jzqh\" (UniqueName: \"kubernetes.io/projected/a0553e25-b483-4ced-898c-6c466cddbdb2-kube-api-access-2jzqh\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.529926 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0553e25-b483-4ced-898c-6c466cddbdb2" (UID: "a0553e25-b483-4ced-898c-6c466cddbdb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.569128 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-config-data" (OuterVolumeSpecName: "config-data") pod "a0553e25-b483-4ced-898c-6c466cddbdb2" (UID: "a0553e25-b483-4ced-898c-6c466cddbdb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.621261 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.621629 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0553e25-b483-4ced-898c-6c466cddbdb2-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.828015 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0553e25-b483-4ced-898c-6c466cddbdb2","Type":"ContainerDied","Data":"483381d5c9a9fe6a79b0bb50834fbdd29e14245b0648b401067cb0d7fd8034b7"} Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.828081 4799 scope.go:117] "RemoveContainer" containerID="654562b9be47c9346b3b506e1f6f79cbf36a6eb2fc613844bdb5f101bebb85e7" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.828302 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.885047 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.894155 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.904836 4799 scope.go:117] "RemoveContainer" containerID="584dbdd8e0cda1ee7f7710b4e068a19d2fc473344da7e9604a9166774c43ed45" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.912411 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:42:42 crc kubenswrapper[4799]: E0930 14:42:42.913423 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="ceilometer-notification-agent" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.913469 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="ceilometer-notification-agent" Sep 30 14:42:42 crc kubenswrapper[4799]: E0930 14:42:42.913564 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="ceilometer-central-agent" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.913577 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="ceilometer-central-agent" Sep 30 14:42:42 crc kubenswrapper[4799]: E0930 14:42:42.913623 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="sg-core" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.913630 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="sg-core" Sep 30 14:42:42 crc kubenswrapper[4799]: E0930 14:42:42.913638 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="proxy-httpd" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.913668 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="proxy-httpd" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.913947 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="ceilometer-notification-agent" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.913987 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="proxy-httpd" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.914004 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="sg-core" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.914019 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" containerName="ceilometer-central-agent" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.919422 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.922891 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.922905 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.969879 4799 scope.go:117] "RemoveContainer" containerID="090b59469bf84d3f8c9547066c5a9c9b3b624194687de7d9b23629a9d56c8aeb" Sep 30 14:42:42 crc kubenswrapper[4799]: I0930 14:42:42.970584 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.009680 4799 scope.go:117] "RemoveContainer" containerID="386c69a6d45ba00c080c113e1e249f3079ecb29291046b293f941e409c9e24b3" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.030566 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-log-httpd\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.030670 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.030827 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.030853 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-scripts\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.031129 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-run-httpd\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.031222 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-config-data\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.031286 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vlvm\" (UniqueName: \"kubernetes.io/projected/6b416afa-3f31-447f-8283-af1bd2ade0fc-kube-api-access-4vlvm\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.138000 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.138085 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-scripts\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.138160 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-run-httpd\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.138211 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-config-data\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.138253 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vlvm\" (UniqueName: \"kubernetes.io/projected/6b416afa-3f31-447f-8283-af1bd2ade0fc-kube-api-access-4vlvm\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.138309 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-log-httpd\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.138348 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.139144 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-run-httpd\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.139359 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-log-httpd\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.142801 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.143571 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-config-data\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.143933 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-scripts\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.157702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.162858 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vlvm\" (UniqueName: \"kubernetes.io/projected/6b416afa-3f31-447f-8283-af1bd2ade0fc-kube-api-access-4vlvm\") pod \"ceilometer-0\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.253362 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.600059 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:42:43 crc kubenswrapper[4799]: I0930 14:42:43.839117 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerStarted","Data":"b30387984614c0ec30b5c08ff4edb7ea67187d499dfb54ce02e0d5cf3f39d282"} Sep 30 14:42:44 crc kubenswrapper[4799]: I0930 14:42:44.519073 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0553e25-b483-4ced-898c-6c466cddbdb2" path="/var/lib/kubelet/pods/a0553e25-b483-4ced-898c-6c466cddbdb2/volumes" Sep 30 14:42:44 crc kubenswrapper[4799]: I0930 14:42:44.851075 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerStarted","Data":"ee11a34e8a354858a54f65100e7546d1995e2abb7f1f4561c6a64358e4595b1c"} Sep 30 14:42:45 crc kubenswrapper[4799]: I0930 14:42:45.878520 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerStarted","Data":"5979ea5bf03171e1964c75cd98fc7c15dfb205529d113bc141f6a5b170af3f6e"} Sep 30 14:42:46 crc kubenswrapper[4799]: I0930 14:42:46.304188 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Sep 30 14:42:46 crc kubenswrapper[4799]: I0930 14:42:46.893377 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerStarted","Data":"b0444f5ffa332872dc4acbe21d5b3e146da1d0ae4f8449809fccef5b55d119f9"} Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.172113 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vwbqw"] Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.173618 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.176452 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.176689 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.194097 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vwbqw"] Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.321327 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-scripts\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.321392 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.321523 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-config-data\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.321545 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc6jl\" (UniqueName: \"kubernetes.io/projected/bdc6f141-7e93-46af-9b59-315ee2bb8025-kube-api-access-kc6jl\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.395399 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.397004 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.402791 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.421735 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.422900 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.423063 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-config-data\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.423088 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc6jl\" (UniqueName: \"kubernetes.io/projected/bdc6f141-7e93-46af-9b59-315ee2bb8025-kube-api-access-kc6jl\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.423140 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-scripts\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.433530 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.437436 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-config-data\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.444581 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-scripts\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.453573 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc6jl\" (UniqueName: \"kubernetes.io/projected/bdc6f141-7e93-46af-9b59-315ee2bb8025-kube-api-access-kc6jl\") pod \"nova-cell0-cell-mapping-vwbqw\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.491389 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.527342 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.527753 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-config-data\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.527899 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5g9g\" (UniqueName: \"kubernetes.io/projected/b0e52c74-757f-473c-be05-25a37c0cad63-kube-api-access-v5g9g\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.544816 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.617959 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.630233 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.632060 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.632110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-config-data\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.632201 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5g9g\" (UniqueName: \"kubernetes.io/projected/b0e52c74-757f-473c-be05-25a37c0cad63-kube-api-access-v5g9g\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.647793 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.654914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.711793 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.713611 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.715195 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-config-data\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.752332 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.801330 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5g9g\" (UniqueName: \"kubernetes.io/projected/b0e52c74-757f-473c-be05-25a37c0cad63-kube-api-access-v5g9g\") pod \"nova-scheduler-0\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " pod="openstack/nova-scheduler-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.822712 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.852342 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.852409 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dwdz\" (UniqueName: \"kubernetes.io/projected/f8294b29-6a0c-4547-ae1f-364ca994faeb-kube-api-access-5dwdz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.852490 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8qdb\" (UniqueName: \"kubernetes.io/projected/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-kube-api-access-t8qdb\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.852627 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.852688 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.852748 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-config-data\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.852813 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-logs\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.903582 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.960065 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8qdb\" (UniqueName: \"kubernetes.io/projected/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-kube-api-access-t8qdb\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.960166 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.960204 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.960269 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-config-data\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.960340 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-logs\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.960432 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.960458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dwdz\" (UniqueName: \"kubernetes.io/projected/f8294b29-6a0c-4547-ae1f-364ca994faeb-kube-api-access-5dwdz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.969073 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-logs\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.987931 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.988012 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.988914 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-config-data\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:47 crc kubenswrapper[4799]: I0930 14:42:47.989458 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.014663 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.064255 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8qdb\" (UniqueName: \"kubernetes.io/projected/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-kube-api-access-t8qdb\") pod \"nova-metadata-0\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " pod="openstack/nova-metadata-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.075550 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dwdz\" (UniqueName: \"kubernetes.io/projected/f8294b29-6a0c-4547-ae1f-364ca994faeb-kube-api-access-5dwdz\") pod \"nova-cell1-novncproxy-0\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.112736 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xv9g7"] Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.114849 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.127027 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xv9g7"] Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.145307 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.147550 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.171837 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.191680 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.195199 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.208214 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.281268 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm85v\" (UniqueName: \"kubernetes.io/projected/21856f94-450c-4c1a-9cf9-7697d7d22a44-kube-api-access-cm85v\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.281329 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.281389 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-config-data\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.281436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.281489 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6hkg\" (UniqueName: \"kubernetes.io/projected/9f0a6422-8114-48d4-8f42-f006d7dd694a-kube-api-access-b6hkg\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.281518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.281577 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-config\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.293024 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.293092 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.293127 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0a6422-8114-48d4-8f42-f006d7dd694a-logs\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.396482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.396834 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-config\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.396928 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.396965 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.396992 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0a6422-8114-48d4-8f42-f006d7dd694a-logs\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.397056 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm85v\" (UniqueName: \"kubernetes.io/projected/21856f94-450c-4c1a-9cf9-7697d7d22a44-kube-api-access-cm85v\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.397099 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.397161 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-config-data\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.397213 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.397282 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6hkg\" (UniqueName: \"kubernetes.io/projected/9f0a6422-8114-48d4-8f42-f006d7dd694a-kube-api-access-b6hkg\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.406846 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0a6422-8114-48d4-8f42-f006d7dd694a-logs\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.407629 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.409429 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-config\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.410005 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.410403 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.422481 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.429578 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.433522 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-config-data\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.439354 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6hkg\" (UniqueName: \"kubernetes.io/projected/9f0a6422-8114-48d4-8f42-f006d7dd694a-kube-api-access-b6hkg\") pod \"nova-api-0\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.446501 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm85v\" (UniqueName: \"kubernetes.io/projected/21856f94-450c-4c1a-9cf9-7697d7d22a44-kube-api-access-cm85v\") pod \"dnsmasq-dns-845d6d6f59-xv9g7\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.459031 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.481522 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:42:48 crc kubenswrapper[4799]: I0930 14:42:48.824314 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vwbqw"] Sep 30 14:42:49 crc kubenswrapper[4799]: I0930 14:42:49.034585 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vwbqw" event={"ID":"bdc6f141-7e93-46af-9b59-315ee2bb8025","Type":"ContainerStarted","Data":"0da4bbcd2346ef4f56471477ec407a1eb3549af459b15418a20ed8a7f4efb931"} Sep 30 14:42:49 crc kubenswrapper[4799]: I0930 14:42:49.164789 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:42:49 crc kubenswrapper[4799]: W0930 14:42:49.232577 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0e52c74_757f_473c_be05_25a37c0cad63.slice/crio-6b9d72e1c2f3bcc722aff84487fec686bcb8fdbce9a3ff735b2bdd9ae0fbbf76 WatchSource:0}: Error finding container 6b9d72e1c2f3bcc722aff84487fec686bcb8fdbce9a3ff735b2bdd9ae0fbbf76: Status 404 returned error can't find the container with id 6b9d72e1c2f3bcc722aff84487fec686bcb8fdbce9a3ff735b2bdd9ae0fbbf76 Sep 30 14:42:49 crc kubenswrapper[4799]: I0930 14:42:49.390892 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:42:49 crc kubenswrapper[4799]: I0930 14:42:49.791380 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xv9g7"] Sep 30 14:42:49 crc kubenswrapper[4799]: I0930 14:42:49.825269 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:42:49 crc kubenswrapper[4799]: I0930 14:42:49.837785 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.123796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerStarted","Data":"e122a9b6ce81c231cea3fbcecb5741d02ff6e0b0378223608e5cda96a1b17d76"} Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.123974 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.132140 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vwbqw" event={"ID":"bdc6f141-7e93-46af-9b59-315ee2bb8025","Type":"ContainerStarted","Data":"9bf9a8c5a0d0d1bd96932149a0bf7bcb60228b667a1eab56f23473d1713d2758"} Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.141529 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0e52c74-757f-473c-be05-25a37c0cad63","Type":"ContainerStarted","Data":"6b9d72e1c2f3bcc722aff84487fec686bcb8fdbce9a3ff735b2bdd9ae0fbbf76"} Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.145680 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f8294b29-6a0c-4547-ae1f-364ca994faeb","Type":"ContainerStarted","Data":"26dbbd3d64eb40e094132e5fdb8a8c27171cb25823d550b5ba1ee71b694bfd08"} Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.158561 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f0a6422-8114-48d4-8f42-f006d7dd694a","Type":"ContainerStarted","Data":"97eb98801871615904d9ead0951239937c89ca2308e890c4b3b8ea11fefdd0f2"} Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.176911 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af1ca710-85b2-40d1-aa3d-e86d848b2b1e","Type":"ContainerStarted","Data":"845c25f513edd21b1d7551fc3910837a2a162386fea1b64432d0099335db9019"} Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.180371 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" event={"ID":"21856f94-450c-4c1a-9cf9-7697d7d22a44","Type":"ContainerStarted","Data":"9dac06f32e2c1ce83297b3a5f21d0f04ef0d9d27a99eaeb0c02b2d76eeca628c"} Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.183784 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.555459786 podStartE2EDuration="8.18375563s" podCreationTimestamp="2025-09-30 14:42:42 +0000 UTC" firstStartedPulling="2025-09-30 14:42:43.619569182 +0000 UTC m=+1385.703169609" lastFinishedPulling="2025-09-30 14:42:48.247865026 +0000 UTC m=+1390.331465453" observedRunningTime="2025-09-30 14:42:50.164312281 +0000 UTC m=+1392.247912718" watchObservedRunningTime="2025-09-30 14:42:50.18375563 +0000 UTC m=+1392.267356057" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.204352 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vwbqw" podStartSLOduration=3.204314811 podStartE2EDuration="3.204314811s" podCreationTimestamp="2025-09-30 14:42:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:42:50.192148935 +0000 UTC m=+1392.275749372" watchObservedRunningTime="2025-09-30 14:42:50.204314811 +0000 UTC m=+1392.287915268" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.359724 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d6gjx"] Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.362467 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.372613 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.372970 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.403872 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d6gjx"] Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.452045 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-config-data\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.452114 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.452207 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-scripts\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.452249 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljzv9\" (UniqueName: \"kubernetes.io/projected/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-kube-api-access-ljzv9\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.555393 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-scripts\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.557039 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljzv9\" (UniqueName: \"kubernetes.io/projected/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-kube-api-access-ljzv9\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.557550 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-config-data\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.557683 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.595096 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-config-data\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.604224 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljzv9\" (UniqueName: \"kubernetes.io/projected/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-kube-api-access-ljzv9\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.611304 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.611953 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-scripts\") pod \"nova-cell1-conductor-db-sync-d6gjx\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:50 crc kubenswrapper[4799]: I0930 14:42:50.851064 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:42:51 crc kubenswrapper[4799]: I0930 14:42:51.238397 4799 generic.go:334] "Generic (PLEG): container finished" podID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerID="4538be714431660e59b8ffe63642f74a14da10ad42592f9022ea6e1f60e5433e" exitCode=0 Sep 30 14:42:51 crc kubenswrapper[4799]: I0930 14:42:51.238492 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" event={"ID":"21856f94-450c-4c1a-9cf9-7697d7d22a44","Type":"ContainerDied","Data":"4538be714431660e59b8ffe63642f74a14da10ad42592f9022ea6e1f60e5433e"} Sep 30 14:42:51 crc kubenswrapper[4799]: I0930 14:42:51.652682 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d6gjx"] Sep 30 14:42:51 crc kubenswrapper[4799]: W0930 14:42:51.685778 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e276d3c_eb12_4fc1_9d76_3ffff9307d4d.slice/crio-2bb8973d84df8f5fa6c79db300fb21e2939a9499a72eb4e8c9c0550ecef47657 WatchSource:0}: Error finding container 2bb8973d84df8f5fa6c79db300fb21e2939a9499a72eb4e8c9c0550ecef47657: Status 404 returned error can't find the container with id 2bb8973d84df8f5fa6c79db300fb21e2939a9499a72eb4e8c9c0550ecef47657 Sep 30 14:42:52 crc kubenswrapper[4799]: I0930 14:42:52.275379 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" event={"ID":"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d","Type":"ContainerStarted","Data":"d7a54ebcdc008ba4e21e63fb3b06dc2352fdd16160a975d87afa427feef45ae2"} Sep 30 14:42:52 crc kubenswrapper[4799]: I0930 14:42:52.275447 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" event={"ID":"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d","Type":"ContainerStarted","Data":"2bb8973d84df8f5fa6c79db300fb21e2939a9499a72eb4e8c9c0550ecef47657"} Sep 30 14:42:52 crc kubenswrapper[4799]: I0930 14:42:52.298433 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" event={"ID":"21856f94-450c-4c1a-9cf9-7697d7d22a44","Type":"ContainerStarted","Data":"b7b735aabdc6ea4e401e3ef0ca463579b572fae1e090a735c4aa413a219bbe81"} Sep 30 14:42:52 crc kubenswrapper[4799]: I0930 14:42:52.299727 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:52 crc kubenswrapper[4799]: I0930 14:42:52.314371 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" podStartSLOduration=2.314337595 podStartE2EDuration="2.314337595s" podCreationTimestamp="2025-09-30 14:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:42:52.302280742 +0000 UTC m=+1394.385881179" watchObservedRunningTime="2025-09-30 14:42:52.314337595 +0000 UTC m=+1394.397938022" Sep 30 14:42:52 crc kubenswrapper[4799]: I0930 14:42:52.337004 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" podStartSLOduration=5.336975246 podStartE2EDuration="5.336975246s" podCreationTimestamp="2025-09-30 14:42:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:42:52.330116946 +0000 UTC m=+1394.413717393" watchObservedRunningTime="2025-09-30 14:42:52.336975246 +0000 UTC m=+1394.420575673" Sep 30 14:42:52 crc kubenswrapper[4799]: I0930 14:42:52.689487 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:42:54 crc kubenswrapper[4799]: I0930 14:42:54.446377 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:42:54 crc kubenswrapper[4799]: I0930 14:42:54.465304 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:42:57 crc kubenswrapper[4799]: I0930 14:42:57.530572 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:42:57 crc kubenswrapper[4799]: I0930 14:42:57.533082 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:42:57 crc kubenswrapper[4799]: I0930 14:42:57.534194 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"6a24eead33c55554d4fa163faed09d5db7e530d41a8e1d35cb03d928a56c1750"} pod="openstack/horizon-7f99689854-glm7t" containerMessage="Container horizon failed startup probe, will be restarted" Sep 30 14:42:57 crc kubenswrapper[4799]: I0930 14:42:57.534248 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" containerID="cri-o://6a24eead33c55554d4fa163faed09d5db7e530d41a8e1d35cb03d928a56c1750" gracePeriod=30 Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.393353 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af1ca710-85b2-40d1-aa3d-e86d848b2b1e","Type":"ContainerStarted","Data":"8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d"} Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.393726 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af1ca710-85b2-40d1-aa3d-e86d848b2b1e","Type":"ContainerStarted","Data":"4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218"} Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.393889 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerName="nova-metadata-log" containerID="cri-o://4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218" gracePeriod=30 Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.394768 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerName="nova-metadata-metadata" containerID="cri-o://8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d" gracePeriod=30 Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.402080 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0e52c74-757f-473c-be05-25a37c0cad63","Type":"ContainerStarted","Data":"2f335d29bd31d7acdbf0fa78ad97b82d7bd911fcf14946e6ab38796f3ba476fe"} Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.405978 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="f8294b29-6a0c-4547-ae1f-364ca994faeb" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://3d508ff494e4d0d11128c3b2bde153940fd37628a81ae86d8c8fabf83a153916" gracePeriod=30 Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.406094 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f8294b29-6a0c-4547-ae1f-364ca994faeb","Type":"ContainerStarted","Data":"3d508ff494e4d0d11128c3b2bde153940fd37628a81ae86d8c8fabf83a153916"} Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.440583 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f0a6422-8114-48d4-8f42-f006d7dd694a","Type":"ContainerStarted","Data":"6809f813c4da12dbdd177859e114418f1aa31b29b9bef4c3dc8eeb0e81158e4b"} Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.440657 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f0a6422-8114-48d4-8f42-f006d7dd694a","Type":"ContainerStarted","Data":"eb6fa930d9371c43bc9dfc67b1f7bb4e762794a7ac4f70611ca374c77cf4b337"} Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.464822 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.470445 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.305068425 podStartE2EDuration="11.47041313s" podCreationTimestamp="2025-09-30 14:42:47 +0000 UTC" firstStartedPulling="2025-09-30 14:42:49.247457518 +0000 UTC m=+1391.331057945" lastFinishedPulling="2025-09-30 14:42:57.412802223 +0000 UTC m=+1399.496402650" observedRunningTime="2025-09-30 14:42:58.462173899 +0000 UTC m=+1400.545774346" watchObservedRunningTime="2025-09-30 14:42:58.47041313 +0000 UTC m=+1400.554013557" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.476725 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.02526162 podStartE2EDuration="11.476694394s" podCreationTimestamp="2025-09-30 14:42:47 +0000 UTC" firstStartedPulling="2025-09-30 14:42:49.812313021 +0000 UTC m=+1391.895913448" lastFinishedPulling="2025-09-30 14:42:57.263745795 +0000 UTC m=+1399.347346222" observedRunningTime="2025-09-30 14:42:58.433951054 +0000 UTC m=+1400.517551491" watchObservedRunningTime="2025-09-30 14:42:58.476694394 +0000 UTC m=+1400.560294841" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.484529 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.484602 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.487145 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": dial tcp 10.217.0.192:8774: connect: connection refused" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.487582 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": dial tcp 10.217.0.192:8774: connect: connection refused" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.494574 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.6406509160000002 podStartE2EDuration="11.494542746s" podCreationTimestamp="2025-09-30 14:42:47 +0000 UTC" firstStartedPulling="2025-09-30 14:42:49.405625572 +0000 UTC m=+1391.489225999" lastFinishedPulling="2025-09-30 14:42:57.259517402 +0000 UTC m=+1399.343117829" observedRunningTime="2025-09-30 14:42:58.490119466 +0000 UTC m=+1400.573719923" watchObservedRunningTime="2025-09-30 14:42:58.494542746 +0000 UTC m=+1400.578143173" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.544257 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.091916558 podStartE2EDuration="11.544230368s" podCreationTimestamp="2025-09-30 14:42:47 +0000 UTC" firstStartedPulling="2025-09-30 14:42:49.804043879 +0000 UTC m=+1391.887644306" lastFinishedPulling="2025-09-30 14:42:57.256357689 +0000 UTC m=+1399.339958116" observedRunningTime="2025-09-30 14:42:58.527066586 +0000 UTC m=+1400.610667023" watchObservedRunningTime="2025-09-30 14:42:58.544230368 +0000 UTC m=+1400.627830795" Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.645089 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jdtxl"] Sep 30 14:42:58 crc kubenswrapper[4799]: I0930 14:42:58.645445 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" podUID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" containerName="dnsmasq-dns" containerID="cri-o://0dbe8d62cbbbfc49cb340f4e23a3440773b4a36fb9035d93f4596e3bb41f36fd" gracePeriod=10 Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.503927 4799 generic.go:334] "Generic (PLEG): container finished" podID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerID="4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218" exitCode=143 Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.504074 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af1ca710-85b2-40d1-aa3d-e86d848b2b1e","Type":"ContainerDied","Data":"4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218"} Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.542897 4799 generic.go:334] "Generic (PLEG): container finished" podID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" containerID="0dbe8d62cbbbfc49cb340f4e23a3440773b4a36fb9035d93f4596e3bb41f36fd" exitCode=0 Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.544045 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" event={"ID":"7720d30e-281a-4f4f-be9c-0b8f1bc475f6","Type":"ContainerDied","Data":"0dbe8d62cbbbfc49cb340f4e23a3440773b4a36fb9035d93f4596e3bb41f36fd"} Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.544087 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" event={"ID":"7720d30e-281a-4f4f-be9c-0b8f1bc475f6","Type":"ContainerDied","Data":"6a3d416cda5549819559da5d14b45d38267f8d91e2ec42e028ddd5525e819079"} Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.544102 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a3d416cda5549819559da5d14b45d38267f8d91e2ec42e028ddd5525e819079" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.571311 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.664462 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-swift-storage-0\") pod \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.664526 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-sb\") pod \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.664634 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljnxj\" (UniqueName: \"kubernetes.io/projected/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-kube-api-access-ljnxj\") pod \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.664731 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-nb\") pod \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.664786 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-config\") pod \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.664843 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-svc\") pod \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\" (UID: \"7720d30e-281a-4f4f-be9c-0b8f1bc475f6\") " Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.713928 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-kube-api-access-ljnxj" (OuterVolumeSpecName: "kube-api-access-ljnxj") pod "7720d30e-281a-4f4f-be9c-0b8f1bc475f6" (UID: "7720d30e-281a-4f4f-be9c-0b8f1bc475f6"). InnerVolumeSpecName "kube-api-access-ljnxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.778884 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljnxj\" (UniqueName: \"kubernetes.io/projected/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-kube-api-access-ljnxj\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.800858 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7720d30e-281a-4f4f-be9c-0b8f1bc475f6" (UID: "7720d30e-281a-4f4f-be9c-0b8f1bc475f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.822969 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-config" (OuterVolumeSpecName: "config") pod "7720d30e-281a-4f4f-be9c-0b8f1bc475f6" (UID: "7720d30e-281a-4f4f-be9c-0b8f1bc475f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.833936 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7720d30e-281a-4f4f-be9c-0b8f1bc475f6" (UID: "7720d30e-281a-4f4f-be9c-0b8f1bc475f6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.841416 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7720d30e-281a-4f4f-be9c-0b8f1bc475f6" (UID: "7720d30e-281a-4f4f-be9c-0b8f1bc475f6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.851821 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7720d30e-281a-4f4f-be9c-0b8f1bc475f6" (UID: "7720d30e-281a-4f4f-be9c-0b8f1bc475f6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.881292 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.881331 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.881341 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.881350 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:42:59 crc kubenswrapper[4799]: I0930 14:42:59.881359 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7720d30e-281a-4f4f-be9c-0b8f1bc475f6-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:00 crc kubenswrapper[4799]: I0930 14:43:00.555550 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-jdtxl" Sep 30 14:43:00 crc kubenswrapper[4799]: I0930 14:43:00.587375 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jdtxl"] Sep 30 14:43:00 crc kubenswrapper[4799]: I0930 14:43:00.614839 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jdtxl"] Sep 30 14:43:02 crc kubenswrapper[4799]: I0930 14:43:02.532703 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" path="/var/lib/kubelet/pods/7720d30e-281a-4f4f-be9c-0b8f1bc475f6/volumes" Sep 30 14:43:02 crc kubenswrapper[4799]: I0930 14:43:02.686968 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:43:02 crc kubenswrapper[4799]: I0930 14:43:02.687093 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:43:02 crc kubenswrapper[4799]: I0930 14:43:02.688141 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"145eff7bbb6dd0a52fa1de14ab2ef3666affe0c5dae478a2644a1c0dffeb1113"} pod="openstack/horizon-7f7c888686-pwv2q" containerMessage="Container horizon failed startup probe, will be restarted" Sep 30 14:43:02 crc kubenswrapper[4799]: I0930 14:43:02.688181 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" containerID="cri-o://145eff7bbb6dd0a52fa1de14ab2ef3666affe0c5dae478a2644a1c0dffeb1113" gracePeriod=30 Sep 30 14:43:03 crc kubenswrapper[4799]: I0930 14:43:03.016452 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 14:43:03 crc kubenswrapper[4799]: I0930 14:43:03.193021 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:43:03 crc kubenswrapper[4799]: I0930 14:43:03.193150 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:43:03 crc kubenswrapper[4799]: I0930 14:43:03.209494 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:05 crc kubenswrapper[4799]: I0930 14:43:05.630337 4799 generic.go:334] "Generic (PLEG): container finished" podID="4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" containerID="d7a54ebcdc008ba4e21e63fb3b06dc2352fdd16160a975d87afa427feef45ae2" exitCode=0 Sep 30 14:43:05 crc kubenswrapper[4799]: I0930 14:43:05.630433 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" event={"ID":"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d","Type":"ContainerDied","Data":"d7a54ebcdc008ba4e21e63fb3b06dc2352fdd16160a975d87afa427feef45ae2"} Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.126268 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.235601 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-combined-ca-bundle\") pod \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.235705 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-config-data\") pod \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.235810 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-scripts\") pod \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.235870 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljzv9\" (UniqueName: \"kubernetes.io/projected/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-kube-api-access-ljzv9\") pod \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\" (UID: \"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d\") " Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.245391 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-kube-api-access-ljzv9" (OuterVolumeSpecName: "kube-api-access-ljzv9") pod "4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" (UID: "4e276d3c-eb12-4fc1-9d76-3ffff9307d4d"). InnerVolumeSpecName "kube-api-access-ljzv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.251027 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-scripts" (OuterVolumeSpecName: "scripts") pod "4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" (UID: "4e276d3c-eb12-4fc1-9d76-3ffff9307d4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.304272 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" (UID: "4e276d3c-eb12-4fc1-9d76-3ffff9307d4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.308110 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-config-data" (OuterVolumeSpecName: "config-data") pod "4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" (UID: "4e276d3c-eb12-4fc1-9d76-3ffff9307d4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.339359 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.339402 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.339414 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.339961 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljzv9\" (UniqueName: \"kubernetes.io/projected/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d-kube-api-access-ljzv9\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.683183 4799 generic.go:334] "Generic (PLEG): container finished" podID="bdc6f141-7e93-46af-9b59-315ee2bb8025" containerID="9bf9a8c5a0d0d1bd96932149a0bf7bcb60228b667a1eab56f23473d1713d2758" exitCode=0 Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.684088 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vwbqw" event={"ID":"bdc6f141-7e93-46af-9b59-315ee2bb8025","Type":"ContainerDied","Data":"9bf9a8c5a0d0d1bd96932149a0bf7bcb60228b667a1eab56f23473d1713d2758"} Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.719048 4799 generic.go:334] "Generic (PLEG): container finished" podID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerID="145eff7bbb6dd0a52fa1de14ab2ef3666affe0c5dae478a2644a1c0dffeb1113" exitCode=0 Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.719195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerDied","Data":"145eff7bbb6dd0a52fa1de14ab2ef3666affe0c5dae478a2644a1c0dffeb1113"} Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.719251 4799 scope.go:117] "RemoveContainer" containerID="2aa2de6ee9c44af7d31b1105ea4177bb893946376162ccd5d747ef4d0d8305e3" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.742139 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" event={"ID":"4e276d3c-eb12-4fc1-9d76-3ffff9307d4d","Type":"ContainerDied","Data":"2bb8973d84df8f5fa6c79db300fb21e2939a9499a72eb4e8c9c0550ecef47657"} Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.742223 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bb8973d84df8f5fa6c79db300fb21e2939a9499a72eb4e8c9c0550ecef47657" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.742374 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-d6gjx" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.863300 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 14:43:07 crc kubenswrapper[4799]: E0930 14:43:07.863769 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" containerName="nova-cell1-conductor-db-sync" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.863796 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" containerName="nova-cell1-conductor-db-sync" Sep 30 14:43:07 crc kubenswrapper[4799]: E0930 14:43:07.863838 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" containerName="init" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.863849 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" containerName="init" Sep 30 14:43:07 crc kubenswrapper[4799]: E0930 14:43:07.863858 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" containerName="dnsmasq-dns" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.863864 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" containerName="dnsmasq-dns" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.864070 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" containerName="nova-cell1-conductor-db-sync" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.864094 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7720d30e-281a-4f4f-be9c-0b8f1bc475f6" containerName="dnsmasq-dns" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.864993 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.870602 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.903147 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.986159 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.986829 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t57lq\" (UniqueName: \"kubernetes.io/projected/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-kube-api-access-t57lq\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:07 crc kubenswrapper[4799]: I0930 14:43:07.986917 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.016308 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.056114 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.092695 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t57lq\" (UniqueName: \"kubernetes.io/projected/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-kube-api-access-t57lq\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.092752 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.092854 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.101593 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.103937 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.141194 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t57lq\" (UniqueName: \"kubernetes.io/projected/ca946c79-58c1-4f6b-9f61-e685bc6dcda9-kube-api-access-t57lq\") pod \"nova-cell1-conductor-0\" (UID: \"ca946c79-58c1-4f6b-9f61-e685bc6dcda9\") " pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.194112 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.782538 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerStarted","Data":"7630e75f64223c31df72c9c0ad92ebccc462f64abfa85db6dbbf91edb9c7433c"} Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.873025 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 14:43:08 crc kubenswrapper[4799]: I0930 14:43:08.904711 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.417692 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x87rk"] Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.424599 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.455516 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.461996 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x87rk"] Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.466896 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-utilities\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.467051 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-catalog-content\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.467078 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktr7p\" (UniqueName: \"kubernetes.io/projected/8f6380d4-aaf5-41ac-b134-9f46269f2183-kube-api-access-ktr7p\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.568711 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-combined-ca-bundle\") pod \"bdc6f141-7e93-46af-9b59-315ee2bb8025\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.569115 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-config-data\") pod \"bdc6f141-7e93-46af-9b59-315ee2bb8025\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.569234 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-scripts\") pod \"bdc6f141-7e93-46af-9b59-315ee2bb8025\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.569374 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc6jl\" (UniqueName: \"kubernetes.io/projected/bdc6f141-7e93-46af-9b59-315ee2bb8025-kube-api-access-kc6jl\") pod \"bdc6f141-7e93-46af-9b59-315ee2bb8025\" (UID: \"bdc6f141-7e93-46af-9b59-315ee2bb8025\") " Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.569982 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-catalog-content\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.570120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktr7p\" (UniqueName: \"kubernetes.io/projected/8f6380d4-aaf5-41ac-b134-9f46269f2183-kube-api-access-ktr7p\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.570394 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-utilities\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.571394 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-utilities\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.576812 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.576886 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.579982 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-catalog-content\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.628718 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdc6f141-7e93-46af-9b59-315ee2bb8025-kube-api-access-kc6jl" (OuterVolumeSpecName: "kube-api-access-kc6jl") pod "bdc6f141-7e93-46af-9b59-315ee2bb8025" (UID: "bdc6f141-7e93-46af-9b59-315ee2bb8025"). InnerVolumeSpecName "kube-api-access-kc6jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.642993 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-scripts" (OuterVolumeSpecName: "scripts") pod "bdc6f141-7e93-46af-9b59-315ee2bb8025" (UID: "bdc6f141-7e93-46af-9b59-315ee2bb8025"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.658399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktr7p\" (UniqueName: \"kubernetes.io/projected/8f6380d4-aaf5-41ac-b134-9f46269f2183-kube-api-access-ktr7p\") pod \"redhat-operators-x87rk\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.667985 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-config-data" (OuterVolumeSpecName: "config-data") pod "bdc6f141-7e93-46af-9b59-315ee2bb8025" (UID: "bdc6f141-7e93-46af-9b59-315ee2bb8025"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.683812 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.683864 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.683890 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc6jl\" (UniqueName: \"kubernetes.io/projected/bdc6f141-7e93-46af-9b59-315ee2bb8025-kube-api-access-kc6jl\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.696512 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdc6f141-7e93-46af-9b59-315ee2bb8025" (UID: "bdc6f141-7e93-46af-9b59-315ee2bb8025"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.776641 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.785602 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdc6f141-7e93-46af-9b59-315ee2bb8025-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.843231 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vwbqw" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.843539 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vwbqw" event={"ID":"bdc6f141-7e93-46af-9b59-315ee2bb8025","Type":"ContainerDied","Data":"0da4bbcd2346ef4f56471477ec407a1eb3549af459b15418a20ed8a7f4efb931"} Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.843751 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0da4bbcd2346ef4f56471477ec407a1eb3549af459b15418a20ed8a7f4efb931" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.860819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca946c79-58c1-4f6b-9f61-e685bc6dcda9","Type":"ContainerStarted","Data":"761806e720a58b98f537b0d5aa77f414529047f75c8ec43f12d59a76231132ae"} Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.860886 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca946c79-58c1-4f6b-9f61-e685bc6dcda9","Type":"ContainerStarted","Data":"1d5030d0fb101c10a7b6dcb129bc722f6c0ca35825fffc165d68aefd45da434c"} Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.862147 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:09 crc kubenswrapper[4799]: I0930 14:43:09.900256 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.900224728 podStartE2EDuration="2.900224728s" podCreationTimestamp="2025-09-30 14:43:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:43:09.886023973 +0000 UTC m=+1411.969624410" watchObservedRunningTime="2025-09-30 14:43:09.900224728 +0000 UTC m=+1411.983825155" Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.169312 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.177975 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-log" containerID="cri-o://eb6fa930d9371c43bc9dfc67b1f7bb4e762794a7ac4f70611ca374c77cf4b337" gracePeriod=30 Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.178856 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-api" containerID="cri-o://6809f813c4da12dbdd177859e114418f1aa31b29b9bef4c3dc8eeb0e81158e4b" gracePeriod=30 Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.564922 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x87rk"] Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.579903 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.896125 4799 generic.go:334] "Generic (PLEG): container finished" podID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerID="00f941859362641f79192a151fc04c974ae808d102269a14feb1bcbcdaa0a92a" exitCode=0 Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.896449 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x87rk" event={"ID":"8f6380d4-aaf5-41ac-b134-9f46269f2183","Type":"ContainerDied","Data":"00f941859362641f79192a151fc04c974ae808d102269a14feb1bcbcdaa0a92a"} Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.896528 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x87rk" event={"ID":"8f6380d4-aaf5-41ac-b134-9f46269f2183","Type":"ContainerStarted","Data":"cedd1bf8ca6758888b0bf2c287ee8122ad8174dc996f2ae43a9efb20e3d99a78"} Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.911428 4799 generic.go:334] "Generic (PLEG): container finished" podID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerID="eb6fa930d9371c43bc9dfc67b1f7bb4e762794a7ac4f70611ca374c77cf4b337" exitCode=143 Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.911505 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f0a6422-8114-48d4-8f42-f006d7dd694a","Type":"ContainerDied","Data":"eb6fa930d9371c43bc9dfc67b1f7bb4e762794a7ac4f70611ca374c77cf4b337"} Sep 30 14:43:10 crc kubenswrapper[4799]: I0930 14:43:10.913202 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b0e52c74-757f-473c-be05-25a37c0cad63" containerName="nova-scheduler-scheduler" containerID="cri-o://2f335d29bd31d7acdbf0fa78ad97b82d7bd911fcf14946e6ab38796f3ba476fe" gracePeriod=30 Sep 30 14:43:12 crc kubenswrapper[4799]: I0930 14:43:12.938028 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x87rk" event={"ID":"8f6380d4-aaf5-41ac-b134-9f46269f2183","Type":"ContainerStarted","Data":"b11ad949030a57b82b99be46ae919d927b149a5f302688076abb0b63cfdf85ff"} Sep 30 14:43:13 crc kubenswrapper[4799]: E0930 14:43:13.023324 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f335d29bd31d7acdbf0fa78ad97b82d7bd911fcf14946e6ab38796f3ba476fe" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:43:13 crc kubenswrapper[4799]: E0930 14:43:13.028133 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f335d29bd31d7acdbf0fa78ad97b82d7bd911fcf14946e6ab38796f3ba476fe" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:43:13 crc kubenswrapper[4799]: E0930 14:43:13.036210 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f335d29bd31d7acdbf0fa78ad97b82d7bd911fcf14946e6ab38796f3ba476fe" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:43:13 crc kubenswrapper[4799]: E0930 14:43:13.036305 4799 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b0e52c74-757f-473c-be05-25a37c0cad63" containerName="nova-scheduler-scheduler" Sep 30 14:43:13 crc kubenswrapper[4799]: I0930 14:43:13.310552 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 14:43:13 crc kubenswrapper[4799]: I0930 14:43:13.960770 4799 generic.go:334] "Generic (PLEG): container finished" podID="b0e52c74-757f-473c-be05-25a37c0cad63" containerID="2f335d29bd31d7acdbf0fa78ad97b82d7bd911fcf14946e6ab38796f3ba476fe" exitCode=0 Sep 30 14:43:13 crc kubenswrapper[4799]: I0930 14:43:13.961017 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0e52c74-757f-473c-be05-25a37c0cad63","Type":"ContainerDied","Data":"2f335d29bd31d7acdbf0fa78ad97b82d7bd911fcf14946e6ab38796f3ba476fe"} Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.394964 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.520462 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-config-data\") pod \"b0e52c74-757f-473c-be05-25a37c0cad63\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.521019 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5g9g\" (UniqueName: \"kubernetes.io/projected/b0e52c74-757f-473c-be05-25a37c0cad63-kube-api-access-v5g9g\") pod \"b0e52c74-757f-473c-be05-25a37c0cad63\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.521214 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-combined-ca-bundle\") pod \"b0e52c74-757f-473c-be05-25a37c0cad63\" (UID: \"b0e52c74-757f-473c-be05-25a37c0cad63\") " Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.593968 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0e52c74-757f-473c-be05-25a37c0cad63-kube-api-access-v5g9g" (OuterVolumeSpecName: "kube-api-access-v5g9g") pod "b0e52c74-757f-473c-be05-25a37c0cad63" (UID: "b0e52c74-757f-473c-be05-25a37c0cad63"). InnerVolumeSpecName "kube-api-access-v5g9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.631467 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5g9g\" (UniqueName: \"kubernetes.io/projected/b0e52c74-757f-473c-be05-25a37c0cad63-kube-api-access-v5g9g\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.647337 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0e52c74-757f-473c-be05-25a37c0cad63" (UID: "b0e52c74-757f-473c-be05-25a37c0cad63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.667360 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-config-data" (OuterVolumeSpecName: "config-data") pod "b0e52c74-757f-473c-be05-25a37c0cad63" (UID: "b0e52c74-757f-473c-be05-25a37c0cad63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.733376 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.733700 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e52c74-757f-473c-be05-25a37c0cad63-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.975963 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0e52c74-757f-473c-be05-25a37c0cad63","Type":"ContainerDied","Data":"6b9d72e1c2f3bcc722aff84487fec686bcb8fdbce9a3ff735b2bdd9ae0fbbf76"} Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.976045 4799 scope.go:117] "RemoveContainer" containerID="2f335d29bd31d7acdbf0fa78ad97b82d7bd911fcf14946e6ab38796f3ba476fe" Sep 30 14:43:14 crc kubenswrapper[4799]: I0930 14:43:14.976251 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.069724 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.102739 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.131572 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:43:15 crc kubenswrapper[4799]: E0930 14:43:15.132236 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc6f141-7e93-46af-9b59-315ee2bb8025" containerName="nova-manage" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.132263 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc6f141-7e93-46af-9b59-315ee2bb8025" containerName="nova-manage" Sep 30 14:43:15 crc kubenswrapper[4799]: E0930 14:43:15.132289 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e52c74-757f-473c-be05-25a37c0cad63" containerName="nova-scheduler-scheduler" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.132300 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e52c74-757f-473c-be05-25a37c0cad63" containerName="nova-scheduler-scheduler" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.132539 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdc6f141-7e93-46af-9b59-315ee2bb8025" containerName="nova-manage" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.132569 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0e52c74-757f-473c-be05-25a37c0cad63" containerName="nova-scheduler-scheduler" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.133493 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.149826 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.176157 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.253988 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.254095 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-config-data\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.254187 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n42fl\" (UniqueName: \"kubernetes.io/projected/d7a4ded6-46db-4145-a518-f9b3dee3843e-kube-api-access-n42fl\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.356420 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n42fl\" (UniqueName: \"kubernetes.io/projected/d7a4ded6-46db-4145-a518-f9b3dee3843e-kube-api-access-n42fl\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.357339 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.357457 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-config-data\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.402545 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.412562 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-config-data\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.413204 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n42fl\" (UniqueName: \"kubernetes.io/projected/d7a4ded6-46db-4145-a518-f9b3dee3843e-kube-api-access-n42fl\") pod \"nova-scheduler-0\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " pod="openstack/nova-scheduler-0" Sep 30 14:43:15 crc kubenswrapper[4799]: I0930 14:43:15.461205 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:43:16 crc kubenswrapper[4799]: I0930 14:43:16.191063 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:43:16 crc kubenswrapper[4799]: I0930 14:43:16.519985 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0e52c74-757f-473c-be05-25a37c0cad63" path="/var/lib/kubelet/pods/b0e52c74-757f-473c-be05-25a37c0cad63/volumes" Sep 30 14:43:17 crc kubenswrapper[4799]: I0930 14:43:17.019842 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7a4ded6-46db-4145-a518-f9b3dee3843e","Type":"ContainerStarted","Data":"21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04"} Sep 30 14:43:17 crc kubenswrapper[4799]: I0930 14:43:17.020012 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7a4ded6-46db-4145-a518-f9b3dee3843e","Type":"ContainerStarted","Data":"9f55b8b6a4452906ff32bcc330a55b8e2842445e86ca741ea7fe3bebed621b70"} Sep 30 14:43:17 crc kubenswrapper[4799]: I0930 14:43:17.681026 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:43:17 crc kubenswrapper[4799]: I0930 14:43:17.681450 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:43:17 crc kubenswrapper[4799]: I0930 14:43:17.684755 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:43:18 crc kubenswrapper[4799]: I0930 14:43:18.251280 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Sep 30 14:43:18 crc kubenswrapper[4799]: I0930 14:43:18.294032 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.293999072 podStartE2EDuration="3.293999072s" podCreationTimestamp="2025-09-30 14:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:43:17.044401011 +0000 UTC m=+1419.128001438" watchObservedRunningTime="2025-09-30 14:43:18.293999072 +0000 UTC m=+1420.377599519" Sep 30 14:43:18 crc kubenswrapper[4799]: I0930 14:43:18.481992 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:43:18 crc kubenswrapper[4799]: I0930 14:43:18.482060 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.062174 4799 generic.go:334] "Generic (PLEG): container finished" podID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerID="b11ad949030a57b82b99be46ae919d927b149a5f302688076abb0b63cfdf85ff" exitCode=0 Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.062260 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x87rk" event={"ID":"8f6380d4-aaf5-41ac-b134-9f46269f2183","Type":"ContainerDied","Data":"b11ad949030a57b82b99be46ae919d927b149a5f302688076abb0b63cfdf85ff"} Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.086245 4799 generic.go:334] "Generic (PLEG): container finished" podID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerID="6809f813c4da12dbdd177859e114418f1aa31b29b9bef4c3dc8eeb0e81158e4b" exitCode=0 Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.086306 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f0a6422-8114-48d4-8f42-f006d7dd694a","Type":"ContainerDied","Data":"6809f813c4da12dbdd177859e114418f1aa31b29b9bef4c3dc8eeb0e81158e4b"} Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.384410 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.463528 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.480635 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0a6422-8114-48d4-8f42-f006d7dd694a-logs\") pod \"9f0a6422-8114-48d4-8f42-f006d7dd694a\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.480761 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-config-data\") pod \"9f0a6422-8114-48d4-8f42-f006d7dd694a\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.480854 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-combined-ca-bundle\") pod \"9f0a6422-8114-48d4-8f42-f006d7dd694a\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.481204 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6hkg\" (UniqueName: \"kubernetes.io/projected/9f0a6422-8114-48d4-8f42-f006d7dd694a-kube-api-access-b6hkg\") pod \"9f0a6422-8114-48d4-8f42-f006d7dd694a\" (UID: \"9f0a6422-8114-48d4-8f42-f006d7dd694a\") " Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.483132 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f0a6422-8114-48d4-8f42-f006d7dd694a-logs" (OuterVolumeSpecName: "logs") pod "9f0a6422-8114-48d4-8f42-f006d7dd694a" (UID: "9f0a6422-8114-48d4-8f42-f006d7dd694a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.492498 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f0a6422-8114-48d4-8f42-f006d7dd694a-kube-api-access-b6hkg" (OuterVolumeSpecName: "kube-api-access-b6hkg") pod "9f0a6422-8114-48d4-8f42-f006d7dd694a" (UID: "9f0a6422-8114-48d4-8f42-f006d7dd694a"). InnerVolumeSpecName "kube-api-access-b6hkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.527289 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f0a6422-8114-48d4-8f42-f006d7dd694a" (UID: "9f0a6422-8114-48d4-8f42-f006d7dd694a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.547121 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-config-data" (OuterVolumeSpecName: "config-data") pod "9f0a6422-8114-48d4-8f42-f006d7dd694a" (UID: "9f0a6422-8114-48d4-8f42-f006d7dd694a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.583713 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6hkg\" (UniqueName: \"kubernetes.io/projected/9f0a6422-8114-48d4-8f42-f006d7dd694a-kube-api-access-b6hkg\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.583767 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f0a6422-8114-48d4-8f42-f006d7dd694a-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.583782 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:20 crc kubenswrapper[4799]: I0930 14:43:20.583795 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f0a6422-8114-48d4-8f42-f006d7dd694a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.113612 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x87rk" event={"ID":"8f6380d4-aaf5-41ac-b134-9f46269f2183","Type":"ContainerStarted","Data":"cb7ef54d9ceb77e889a553ae7f46f32011c58b4d366a7258f2be9bc982883bcd"} Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.129509 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f0a6422-8114-48d4-8f42-f006d7dd694a","Type":"ContainerDied","Data":"97eb98801871615904d9ead0951239937c89ca2308e890c4b3b8ea11fefdd0f2"} Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.129586 4799 scope.go:117] "RemoveContainer" containerID="6809f813c4da12dbdd177859e114418f1aa31b29b9bef4c3dc8eeb0e81158e4b" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.129617 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.166104 4799 scope.go:117] "RemoveContainer" containerID="eb6fa930d9371c43bc9dfc67b1f7bb4e762794a7ac4f70611ca374c77cf4b337" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.168625 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x87rk" podStartSLOduration=2.2985929880000002 podStartE2EDuration="12.168609918s" podCreationTimestamp="2025-09-30 14:43:09 +0000 UTC" firstStartedPulling="2025-09-30 14:43:10.898594264 +0000 UTC m=+1412.982194691" lastFinishedPulling="2025-09-30 14:43:20.768611194 +0000 UTC m=+1422.852211621" observedRunningTime="2025-09-30 14:43:21.165676392 +0000 UTC m=+1423.249276809" watchObservedRunningTime="2025-09-30 14:43:21.168609918 +0000 UTC m=+1423.252210345" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.209744 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.236769 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.254673 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:21 crc kubenswrapper[4799]: E0930 14:43:21.255252 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-log" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.255270 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-log" Sep 30 14:43:21 crc kubenswrapper[4799]: E0930 14:43:21.255305 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-api" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.255311 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-api" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.255530 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-log" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.255554 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" containerName="nova-api-api" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.257107 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.261949 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.270072 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.310939 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-config-data\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.311028 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.311069 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z49pk\" (UniqueName: \"kubernetes.io/projected/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-kube-api-access-z49pk\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.311482 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-logs\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.414064 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-logs\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.415183 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-config-data\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.415301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.416876 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z49pk\" (UniqueName: \"kubernetes.io/projected/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-kube-api-access-z49pk\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.417200 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-logs\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.425294 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.427709 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-config-data\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.456977 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z49pk\" (UniqueName: \"kubernetes.io/projected/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-kube-api-access-z49pk\") pod \"nova-api-0\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.611604 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.710262 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:43:21 crc kubenswrapper[4799]: I0930 14:43:21.710577 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="ae8631cb-f4a0-4d95-afb4-64129fc0475a" containerName="kube-state-metrics" containerID="cri-o://0bec0a419315cfe668fcfe77877d66678738b2af950136f5bd931bf40eed6749" gracePeriod=30 Sep 30 14:43:22 crc kubenswrapper[4799]: I0930 14:43:22.149276 4799 generic.go:334] "Generic (PLEG): container finished" podID="ae8631cb-f4a0-4d95-afb4-64129fc0475a" containerID="0bec0a419315cfe668fcfe77877d66678738b2af950136f5bd931bf40eed6749" exitCode=2 Sep 30 14:43:22 crc kubenswrapper[4799]: I0930 14:43:22.149341 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ae8631cb-f4a0-4d95-afb4-64129fc0475a","Type":"ContainerDied","Data":"0bec0a419315cfe668fcfe77877d66678738b2af950136f5bd931bf40eed6749"} Sep 30 14:43:22 crc kubenswrapper[4799]: I0930 14:43:22.342670 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:22 crc kubenswrapper[4799]: I0930 14:43:22.553259 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f0a6422-8114-48d4-8f42-f006d7dd694a" path="/var/lib/kubelet/pods/9f0a6422-8114-48d4-8f42-f006d7dd694a/volumes" Sep 30 14:43:22 crc kubenswrapper[4799]: I0930 14:43:22.615283 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:43:22 crc kubenswrapper[4799]: I0930 14:43:22.666567 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9g7f\" (UniqueName: \"kubernetes.io/projected/ae8631cb-f4a0-4d95-afb4-64129fc0475a-kube-api-access-g9g7f\") pod \"ae8631cb-f4a0-4d95-afb4-64129fc0475a\" (UID: \"ae8631cb-f4a0-4d95-afb4-64129fc0475a\") " Sep 30 14:43:22 crc kubenswrapper[4799]: I0930 14:43:22.702135 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8631cb-f4a0-4d95-afb4-64129fc0475a-kube-api-access-g9g7f" (OuterVolumeSpecName: "kube-api-access-g9g7f") pod "ae8631cb-f4a0-4d95-afb4-64129fc0475a" (UID: "ae8631cb-f4a0-4d95-afb4-64129fc0475a"). InnerVolumeSpecName "kube-api-access-g9g7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:22 crc kubenswrapper[4799]: I0930 14:43:22.769080 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9g7f\" (UniqueName: \"kubernetes.io/projected/ae8631cb-f4a0-4d95-afb4-64129fc0475a-kube-api-access-g9g7f\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.177604 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8","Type":"ContainerStarted","Data":"d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4"} Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.179932 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8","Type":"ContainerStarted","Data":"24b9134b5558104dd356b8966b050fc930230722e072f62034a03cde42eed265"} Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.183875 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ae8631cb-f4a0-4d95-afb4-64129fc0475a","Type":"ContainerDied","Data":"4c288664139750daa91b7f91c69e7f62e450cc61dada8c6ec3718c7a76d19dd6"} Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.183931 4799 scope.go:117] "RemoveContainer" containerID="0bec0a419315cfe668fcfe77877d66678738b2af950136f5bd931bf40eed6749" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.184098 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.282631 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.299268 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.310186 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:43:23 crc kubenswrapper[4799]: E0930 14:43:23.310824 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8631cb-f4a0-4d95-afb4-64129fc0475a" containerName="kube-state-metrics" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.310855 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8631cb-f4a0-4d95-afb4-64129fc0475a" containerName="kube-state-metrics" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.311155 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8631cb-f4a0-4d95-afb4-64129fc0475a" containerName="kube-state-metrics" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.312199 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.317407 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.317749 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.327585 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.384166 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.384299 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd6w9\" (UniqueName: \"kubernetes.io/projected/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-api-access-vd6w9\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.384346 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.384411 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.486762 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.486945 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.489212 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd6w9\" (UniqueName: \"kubernetes.io/projected/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-api-access-vd6w9\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.489703 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.492919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.494431 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.499482 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.529764 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd6w9\" (UniqueName: \"kubernetes.io/projected/b4ecfc5d-9019-459f-83e8-8affcb145e78-kube-api-access-vd6w9\") pod \"kube-state-metrics-0\" (UID: \"b4ecfc5d-9019-459f-83e8-8affcb145e78\") " pod="openstack/kube-state-metrics-0" Sep 30 14:43:23 crc kubenswrapper[4799]: I0930 14:43:23.635199 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Sep 30 14:43:24 crc kubenswrapper[4799]: I0930 14:43:24.200753 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8","Type":"ContainerStarted","Data":"144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c"} Sep 30 14:43:24 crc kubenswrapper[4799]: I0930 14:43:24.223902 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Sep 30 14:43:24 crc kubenswrapper[4799]: W0930 14:43:24.233982 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4ecfc5d_9019_459f_83e8_8affcb145e78.slice/crio-40918265c0a7f5f5a2cbd892b17f0d5ed81aff5a3e396c091c66e82a3cf0dd70 WatchSource:0}: Error finding container 40918265c0a7f5f5a2cbd892b17f0d5ed81aff5a3e396c091c66e82a3cf0dd70: Status 404 returned error can't find the container with id 40918265c0a7f5f5a2cbd892b17f0d5ed81aff5a3e396c091c66e82a3cf0dd70 Sep 30 14:43:24 crc kubenswrapper[4799]: I0930 14:43:24.248693 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.248644459 podStartE2EDuration="3.248644459s" podCreationTimestamp="2025-09-30 14:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:43:24.236112723 +0000 UTC m=+1426.319713160" watchObservedRunningTime="2025-09-30 14:43:24.248644459 +0000 UTC m=+1426.332244896" Sep 30 14:43:24 crc kubenswrapper[4799]: I0930 14:43:24.525562 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8631cb-f4a0-4d95-afb4-64129fc0475a" path="/var/lib/kubelet/pods/ae8631cb-f4a0-4d95-afb4-64129fc0475a/volumes" Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.213824 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b4ecfc5d-9019-459f-83e8-8affcb145e78","Type":"ContainerStarted","Data":"8cb0733b2708c1d0a7bde54e99542234219868f8ead3d2122a30684e81656255"} Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.213894 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b4ecfc5d-9019-459f-83e8-8affcb145e78","Type":"ContainerStarted","Data":"40918265c0a7f5f5a2cbd892b17f0d5ed81aff5a3e396c091c66e82a3cf0dd70"} Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.214002 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.242405 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.86475261 podStartE2EDuration="2.2423739s" podCreationTimestamp="2025-09-30 14:43:23 +0000 UTC" firstStartedPulling="2025-09-30 14:43:24.240432419 +0000 UTC m=+1426.324032846" lastFinishedPulling="2025-09-30 14:43:24.618053709 +0000 UTC m=+1426.701654136" observedRunningTime="2025-09-30 14:43:25.238314421 +0000 UTC m=+1427.321914868" watchObservedRunningTime="2025-09-30 14:43:25.2423739 +0000 UTC m=+1427.325974337" Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.462892 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.517607 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.968135 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.968502 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="ceilometer-central-agent" containerID="cri-o://ee11a34e8a354858a54f65100e7546d1995e2abb7f1f4561c6a64358e4595b1c" gracePeriod=30 Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.968665 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="proxy-httpd" containerID="cri-o://e122a9b6ce81c231cea3fbcecb5741d02ff6e0b0378223608e5cda96a1b17d76" gracePeriod=30 Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.968710 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="ceilometer-notification-agent" containerID="cri-o://5979ea5bf03171e1964c75cd98fc7c15dfb205529d113bc141f6a5b170af3f6e" gracePeriod=30 Sep 30 14:43:25 crc kubenswrapper[4799]: I0930 14:43:25.968791 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="sg-core" containerID="cri-o://b0444f5ffa332872dc4acbe21d5b3e146da1d0ae4f8449809fccef5b55d119f9" gracePeriod=30 Sep 30 14:43:26 crc kubenswrapper[4799]: I0930 14:43:26.245102 4799 generic.go:334] "Generic (PLEG): container finished" podID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerID="e122a9b6ce81c231cea3fbcecb5741d02ff6e0b0378223608e5cda96a1b17d76" exitCode=0 Sep 30 14:43:26 crc kubenswrapper[4799]: I0930 14:43:26.245154 4799 generic.go:334] "Generic (PLEG): container finished" podID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerID="b0444f5ffa332872dc4acbe21d5b3e146da1d0ae4f8449809fccef5b55d119f9" exitCode=2 Sep 30 14:43:26 crc kubenswrapper[4799]: I0930 14:43:26.245219 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerDied","Data":"e122a9b6ce81c231cea3fbcecb5741d02ff6e0b0378223608e5cda96a1b17d76"} Sep 30 14:43:26 crc kubenswrapper[4799]: I0930 14:43:26.245286 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerDied","Data":"b0444f5ffa332872dc4acbe21d5b3e146da1d0ae4f8449809fccef5b55d119f9"} Sep 30 14:43:26 crc kubenswrapper[4799]: I0930 14:43:26.281293 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 14:43:27 crc kubenswrapper[4799]: I0930 14:43:27.259172 4799 generic.go:334] "Generic (PLEG): container finished" podID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerID="ee11a34e8a354858a54f65100e7546d1995e2abb7f1f4561c6a64358e4595b1c" exitCode=0 Sep 30 14:43:27 crc kubenswrapper[4799]: I0930 14:43:27.260365 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerDied","Data":"ee11a34e8a354858a54f65100e7546d1995e2abb7f1f4561c6a64358e4595b1c"} Sep 30 14:43:27 crc kubenswrapper[4799]: I0930 14:43:27.682434 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.278082 4799 generic.go:334] "Generic (PLEG): container finished" podID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerID="5979ea5bf03171e1964c75cd98fc7c15dfb205529d113bc141f6a5b170af3f6e" exitCode=0 Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.278455 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerDied","Data":"5979ea5bf03171e1964c75cd98fc7c15dfb205529d113bc141f6a5b170af3f6e"} Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.278497 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6b416afa-3f31-447f-8283-af1bd2ade0fc","Type":"ContainerDied","Data":"b30387984614c0ec30b5c08ff4edb7ea67187d499dfb54ce02e0d5cf3f39d282"} Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.278515 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b30387984614c0ec30b5c08ff4edb7ea67187d499dfb54ce02e0d5cf3f39d282" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.286772 4799 generic.go:334] "Generic (PLEG): container finished" podID="d15890bb-d088-4126-8382-b5d8e423b608" containerID="6a24eead33c55554d4fa163faed09d5db7e530d41a8e1d35cb03d928a56c1750" exitCode=137 Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.286848 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerDied","Data":"6a24eead33c55554d4fa163faed09d5db7e530d41a8e1d35cb03d928a56c1750"} Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.286889 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerStarted","Data":"e36dbd20d5d3cdf7ece15c8b2611df5884ae476a6712a79abbb5e85a1de0277d"} Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.286914 4799 scope.go:117] "RemoveContainer" containerID="2e60bf33011afb729f10179f6e079df29dfb26c28596d7ab55738fe40269d6c7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.326075 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.433482 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-run-httpd\") pod \"6b416afa-3f31-447f-8283-af1bd2ade0fc\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.433730 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-config-data\") pod \"6b416afa-3f31-447f-8283-af1bd2ade0fc\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.433770 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-log-httpd\") pod \"6b416afa-3f31-447f-8283-af1bd2ade0fc\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.433857 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-scripts\") pod \"6b416afa-3f31-447f-8283-af1bd2ade0fc\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.433930 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vlvm\" (UniqueName: \"kubernetes.io/projected/6b416afa-3f31-447f-8283-af1bd2ade0fc-kube-api-access-4vlvm\") pod \"6b416afa-3f31-447f-8283-af1bd2ade0fc\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.434575 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-combined-ca-bundle\") pod \"6b416afa-3f31-447f-8283-af1bd2ade0fc\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.434665 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-sg-core-conf-yaml\") pod \"6b416afa-3f31-447f-8283-af1bd2ade0fc\" (UID: \"6b416afa-3f31-447f-8283-af1bd2ade0fc\") " Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.434828 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6b416afa-3f31-447f-8283-af1bd2ade0fc" (UID: "6b416afa-3f31-447f-8283-af1bd2ade0fc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.436326 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6b416afa-3f31-447f-8283-af1bd2ade0fc" (UID: "6b416afa-3f31-447f-8283-af1bd2ade0fc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.448409 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.448512 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b416afa-3f31-447f-8283-af1bd2ade0fc-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.482550 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-scripts" (OuterVolumeSpecName: "scripts") pod "6b416afa-3f31-447f-8283-af1bd2ade0fc" (UID: "6b416afa-3f31-447f-8283-af1bd2ade0fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.483200 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b416afa-3f31-447f-8283-af1bd2ade0fc-kube-api-access-4vlvm" (OuterVolumeSpecName: "kube-api-access-4vlvm") pod "6b416afa-3f31-447f-8283-af1bd2ade0fc" (UID: "6b416afa-3f31-447f-8283-af1bd2ade0fc"). InnerVolumeSpecName "kube-api-access-4vlvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.562638 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.562694 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vlvm\" (UniqueName: \"kubernetes.io/projected/6b416afa-3f31-447f-8283-af1bd2ade0fc-kube-api-access-4vlvm\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.587282 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6b416afa-3f31-447f-8283-af1bd2ade0fc" (UID: "6b416afa-3f31-447f-8283-af1bd2ade0fc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.664869 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.665224 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b416afa-3f31-447f-8283-af1bd2ade0fc" (UID: "6b416afa-3f31-447f-8283-af1bd2ade0fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.666111 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-config-data" (OuterVolumeSpecName: "config-data") pod "6b416afa-3f31-447f-8283-af1bd2ade0fc" (UID: "6b416afa-3f31-447f-8283-af1bd2ade0fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.737945 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s72t7"] Sep 30 14:43:28 crc kubenswrapper[4799]: E0930 14:43:28.738531 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="sg-core" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.738548 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="sg-core" Sep 30 14:43:28 crc kubenswrapper[4799]: E0930 14:43:28.738573 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="ceilometer-central-agent" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.738581 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="ceilometer-central-agent" Sep 30 14:43:28 crc kubenswrapper[4799]: E0930 14:43:28.738622 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="ceilometer-notification-agent" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.738635 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="ceilometer-notification-agent" Sep 30 14:43:28 crc kubenswrapper[4799]: E0930 14:43:28.738676 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="proxy-httpd" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.738688 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="proxy-httpd" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.738972 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="proxy-httpd" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.738999 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="ceilometer-central-agent" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.739007 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="sg-core" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.739026 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" containerName="ceilometer-notification-agent" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.741081 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.759936 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s72t7"] Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.766331 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-utilities\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.766770 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q9m2\" (UniqueName: \"kubernetes.io/projected/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-kube-api-access-6q9m2\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.767070 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-catalog-content\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.767325 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.767409 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b416afa-3f31-447f-8283-af1bd2ade0fc-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.869189 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q9m2\" (UniqueName: \"kubernetes.io/projected/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-kube-api-access-6q9m2\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.869299 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-catalog-content\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.869398 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-utilities\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.869950 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-utilities\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.870502 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-catalog-content\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:28 crc kubenswrapper[4799]: I0930 14:43:28.933388 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q9m2\" (UniqueName: \"kubernetes.io/projected/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-kube-api-access-6q9m2\") pod \"community-operators-s72t7\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.115635 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.301352 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.326358 4799 generic.go:334] "Generic (PLEG): container finished" podID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerID="8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d" exitCode=137 Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.326479 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af1ca710-85b2-40d1-aa3d-e86d848b2b1e","Type":"ContainerDied","Data":"8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d"} Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.326528 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"af1ca710-85b2-40d1-aa3d-e86d848b2b1e","Type":"ContainerDied","Data":"845c25f513edd21b1d7551fc3910837a2a162386fea1b64432d0099335db9019"} Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.326549 4799 scope.go:117] "RemoveContainer" containerID="8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.326785 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.336605 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.399115 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-config-data\") pod \"f8294b29-6a0c-4547-ae1f-364ca994faeb\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.399183 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-logs\") pod \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.399223 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-combined-ca-bundle\") pod \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.399355 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-combined-ca-bundle\") pod \"f8294b29-6a0c-4547-ae1f-364ca994faeb\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.399426 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-config-data\") pod \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.399521 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dwdz\" (UniqueName: \"kubernetes.io/projected/f8294b29-6a0c-4547-ae1f-364ca994faeb-kube-api-access-5dwdz\") pod \"f8294b29-6a0c-4547-ae1f-364ca994faeb\" (UID: \"f8294b29-6a0c-4547-ae1f-364ca994faeb\") " Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.399556 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8qdb\" (UniqueName: \"kubernetes.io/projected/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-kube-api-access-t8qdb\") pod \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\" (UID: \"af1ca710-85b2-40d1-aa3d-e86d848b2b1e\") " Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.400427 4799 generic.go:334] "Generic (PLEG): container finished" podID="f8294b29-6a0c-4547-ae1f-364ca994faeb" containerID="3d508ff494e4d0d11128c3b2bde153940fd37628a81ae86d8c8fabf83a153916" exitCode=137 Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.400565 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.401780 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.401869 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f8294b29-6a0c-4547-ae1f-364ca994faeb","Type":"ContainerDied","Data":"3d508ff494e4d0d11128c3b2bde153940fd37628a81ae86d8c8fabf83a153916"} Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.410155 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-kube-api-access-t8qdb" (OuterVolumeSpecName: "kube-api-access-t8qdb") pod "af1ca710-85b2-40d1-aa3d-e86d848b2b1e" (UID: "af1ca710-85b2-40d1-aa3d-e86d848b2b1e"). InnerVolumeSpecName "kube-api-access-t8qdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.421740 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-logs" (OuterVolumeSpecName: "logs") pod "af1ca710-85b2-40d1-aa3d-e86d848b2b1e" (UID: "af1ca710-85b2-40d1-aa3d-e86d848b2b1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.438086 4799 scope.go:117] "RemoveContainer" containerID="4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.438414 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8294b29-6a0c-4547-ae1f-364ca994faeb-kube-api-access-5dwdz" (OuterVolumeSpecName: "kube-api-access-5dwdz") pod "f8294b29-6a0c-4547-ae1f-364ca994faeb" (UID: "f8294b29-6a0c-4547-ae1f-364ca994faeb"). InnerVolumeSpecName "kube-api-access-5dwdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.519091 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-config-data" (OuterVolumeSpecName: "config-data") pod "f8294b29-6a0c-4547-ae1f-364ca994faeb" (UID: "f8294b29-6a0c-4547-ae1f-364ca994faeb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.532342 4799 scope.go:117] "RemoveContainer" containerID="8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.534470 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dwdz\" (UniqueName: \"kubernetes.io/projected/f8294b29-6a0c-4547-ae1f-364ca994faeb-kube-api-access-5dwdz\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.534514 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8qdb\" (UniqueName: \"kubernetes.io/projected/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-kube-api-access-t8qdb\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.534526 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.534537 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:29 crc kubenswrapper[4799]: E0930 14:43:29.534748 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d\": container with ID starting with 8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d not found: ID does not exist" containerID="8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.534797 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d"} err="failed to get container status \"8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d\": rpc error: code = NotFound desc = could not find container \"8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d\": container with ID starting with 8af86d3a3103d6ff070c88d7b8858706225182c1079f5cd0a293f2776f8db03d not found: ID does not exist" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.534828 4799 scope.go:117] "RemoveContainer" containerID="4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.534899 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8294b29-6a0c-4547-ae1f-364ca994faeb" (UID: "f8294b29-6a0c-4547-ae1f-364ca994faeb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:29 crc kubenswrapper[4799]: E0930 14:43:29.536859 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218\": container with ID starting with 4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218 not found: ID does not exist" containerID="4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.536892 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218"} err="failed to get container status \"4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218\": rpc error: code = NotFound desc = could not find container \"4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218\": container with ID starting with 4cbe9f7cc9bc80e90d41fb0542c8b5cc767c49052e38b32f2a405a311f028218 not found: ID does not exist" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.536911 4799 scope.go:117] "RemoveContainer" containerID="3d508ff494e4d0d11128c3b2bde153940fd37628a81ae86d8c8fabf83a153916" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.576995 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.610577 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.611944 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af1ca710-85b2-40d1-aa3d-e86d848b2b1e" (UID: "af1ca710-85b2-40d1-aa3d-e86d848b2b1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.615857 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-config-data" (OuterVolumeSpecName: "config-data") pod "af1ca710-85b2-40d1-aa3d-e86d848b2b1e" (UID: "af1ca710-85b2-40d1-aa3d-e86d848b2b1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.627155 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:29 crc kubenswrapper[4799]: E0930 14:43:29.627755 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerName="nova-metadata-metadata" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.627781 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerName="nova-metadata-metadata" Sep 30 14:43:29 crc kubenswrapper[4799]: E0930 14:43:29.627805 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8294b29-6a0c-4547-ae1f-364ca994faeb" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.627814 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8294b29-6a0c-4547-ae1f-364ca994faeb" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 14:43:29 crc kubenswrapper[4799]: E0930 14:43:29.627824 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerName="nova-metadata-log" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.627832 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerName="nova-metadata-log" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.628056 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8294b29-6a0c-4547-ae1f-364ca994faeb" containerName="nova-cell1-novncproxy-novncproxy" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.628096 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerName="nova-metadata-metadata" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.628118 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" containerName="nova-metadata-log" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.631220 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.637642 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8294b29-6a0c-4547-ae1f-364ca994faeb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.637699 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.637712 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1ca710-85b2-40d1-aa3d-e86d848b2b1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.642104 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.642481 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.642684 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.744474 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-run-httpd\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.744532 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-config-data\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.744563 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-scripts\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.744581 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-log-httpd\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.744600 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.744702 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.744978 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.745023 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g2rw\" (UniqueName: \"kubernetes.io/projected/5035aa23-c338-4ae2-8de9-94ac166052c9-kube-api-access-2g2rw\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.767575 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.777905 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.777972 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.845121 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.848241 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.848385 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.848447 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g2rw\" (UniqueName: \"kubernetes.io/projected/5035aa23-c338-4ae2-8de9-94ac166052c9-kube-api-access-2g2rw\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.848526 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-run-httpd\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.848551 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-config-data\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.848576 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-scripts\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.848591 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-log-httpd\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.848608 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.859129 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-log-httpd\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.859567 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-run-httpd\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.871074 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-scripts\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.871169 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.895888 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.896374 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-config-data\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.896605 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.909119 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.912515 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g2rw\" (UniqueName: \"kubernetes.io/projected/5035aa23-c338-4ae2-8de9-94ac166052c9-kube-api-access-2g2rw\") pod \"ceilometer-0\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " pod="openstack/ceilometer-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.943769 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.946315 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.959267 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.959524 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 14:43:29 crc kubenswrapper[4799]: I0930 14:43:29.996473 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.017746 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.033328 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.049299 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.052585 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.055520 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-config-data\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.061193 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqdrt\" (UniqueName: \"kubernetes.io/projected/ce901e13-dfa1-4b79-a584-4903008803d0-kube-api-access-zqdrt\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.061670 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.057432 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.061948 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.057513 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.062293 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce901e13-dfa1-4b79-a584-4903008803d0-logs\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.058209 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.064810 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.105947 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.164573 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s72t7"] Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.172425 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqdrt\" (UniqueName: \"kubernetes.io/projected/ce901e13-dfa1-4b79-a584-4903008803d0-kube-api-access-zqdrt\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.172965 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.173151 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.173892 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.177600 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce901e13-dfa1-4b79-a584-4903008803d0-logs\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.177754 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.177826 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.182592 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.182722 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-config-data\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.182807 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk2t8\" (UniqueName: \"kubernetes.io/projected/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-kube-api-access-vk2t8\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.184904 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce901e13-dfa1-4b79-a584-4903008803d0-logs\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.204607 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.216427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.234474 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-config-data\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.244604 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqdrt\" (UniqueName: \"kubernetes.io/projected/ce901e13-dfa1-4b79-a584-4903008803d0-kube-api-access-zqdrt\") pod \"nova-metadata-0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.286434 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk2t8\" (UniqueName: \"kubernetes.io/projected/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-kube-api-access-vk2t8\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.286915 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.287866 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.287951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.288126 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.299069 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.302759 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.305479 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.340881 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.345489 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.348330 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk2t8\" (UniqueName: \"kubernetes.io/projected/a5fd7be8-ed88-4d3c-b979-8d944ca7b864-kube-api-access-vk2t8\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5fd7be8-ed88-4d3c-b979-8d944ca7b864\") " pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.394025 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.455525 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s72t7" event={"ID":"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f","Type":"ContainerStarted","Data":"39c99460ce9195bd3d8a63eef52e21b38f6334e37370ea8d91f9f0cc7c8793c0"} Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.662251 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b416afa-3f31-447f-8283-af1bd2ade0fc" path="/var/lib/kubelet/pods/6b416afa-3f31-447f-8283-af1bd2ade0fc/volumes" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.673036 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af1ca710-85b2-40d1-aa3d-e86d848b2b1e" path="/var/lib/kubelet/pods/af1ca710-85b2-40d1-aa3d-e86d848b2b1e/volumes" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.674253 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8294b29-6a0c-4547-ae1f-364ca994faeb" path="/var/lib/kubelet/pods/f8294b29-6a0c-4547-ae1f-364ca994faeb/volumes" Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.922528 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x87rk" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="registry-server" probeResult="failure" output=< Sep 30 14:43:30 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 14:43:30 crc kubenswrapper[4799]: > Sep 30 14:43:30 crc kubenswrapper[4799]: I0930 14:43:30.950618 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:31 crc kubenswrapper[4799]: I0930 14:43:31.346009 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Sep 30 14:43:31 crc kubenswrapper[4799]: I0930 14:43:31.548917 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:43:31 crc kubenswrapper[4799]: I0930 14:43:31.558585 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a5fd7be8-ed88-4d3c-b979-8d944ca7b864","Type":"ContainerStarted","Data":"4e9f439f789a096133b6ac4b5f2c6d509181079ce97def7f9b6dd46ba0c5b7d2"} Sep 30 14:43:31 crc kubenswrapper[4799]: W0930 14:43:31.562079 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce901e13_dfa1_4b79_a584_4903008803d0.slice/crio-eb27a9bd03bfd7cd7fdda4caec0b0117181b206f4f17304fd413ae0edf44237b WatchSource:0}: Error finding container eb27a9bd03bfd7cd7fdda4caec0b0117181b206f4f17304fd413ae0edf44237b: Status 404 returned error can't find the container with id eb27a9bd03bfd7cd7fdda4caec0b0117181b206f4f17304fd413ae0edf44237b Sep 30 14:43:31 crc kubenswrapper[4799]: I0930 14:43:31.564984 4799 generic.go:334] "Generic (PLEG): container finished" podID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerID="7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4" exitCode=0 Sep 30 14:43:31 crc kubenswrapper[4799]: I0930 14:43:31.566380 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s72t7" event={"ID":"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f","Type":"ContainerDied","Data":"7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4"} Sep 30 14:43:31 crc kubenswrapper[4799]: I0930 14:43:31.585934 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerStarted","Data":"fc54f687230fc12b273b42d1f7a4f8486a65f34d54d3735917b86b4436fe689d"} Sep 30 14:43:31 crc kubenswrapper[4799]: I0930 14:43:31.632931 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:43:31 crc kubenswrapper[4799]: I0930 14:43:31.639879 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.631101 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce901e13-dfa1-4b79-a584-4903008803d0","Type":"ContainerStarted","Data":"fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44"} Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.631516 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce901e13-dfa1-4b79-a584-4903008803d0","Type":"ContainerStarted","Data":"d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7"} Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.631557 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce901e13-dfa1-4b79-a584-4903008803d0","Type":"ContainerStarted","Data":"eb27a9bd03bfd7cd7fdda4caec0b0117181b206f4f17304fd413ae0edf44237b"} Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.653530 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a5fd7be8-ed88-4d3c-b979-8d944ca7b864","Type":"ContainerStarted","Data":"e36e7474482177a6c9b756afded8e9dd48eaa3b7d6552dc9544bf184b7edc706"} Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.677217 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerStarted","Data":"c46f8ed55b6f42fce896e3a0194f62a1a70b9de0090d8d0e77cee93983e2bdfd"} Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.677982 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.677970201 podStartE2EDuration="3.677970201s" podCreationTimestamp="2025-09-30 14:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:43:32.668316609 +0000 UTC m=+1434.751917046" watchObservedRunningTime="2025-09-30 14:43:32.677970201 +0000 UTC m=+1434.761570628" Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.719824 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.720439 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:43:32 crc kubenswrapper[4799]: I0930 14:43:32.724739 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.724711328 podStartE2EDuration="3.724711328s" podCreationTimestamp="2025-09-30 14:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:43:32.702311973 +0000 UTC m=+1434.785912420" watchObservedRunningTime="2025-09-30 14:43:32.724711328 +0000 UTC m=+1434.808311765" Sep 30 14:43:33 crc kubenswrapper[4799]: I0930 14:43:33.700769 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Sep 30 14:43:33 crc kubenswrapper[4799]: I0930 14:43:33.773093 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s72t7" event={"ID":"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f","Type":"ContainerStarted","Data":"75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b"} Sep 30 14:43:33 crc kubenswrapper[4799]: I0930 14:43:33.781795 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerStarted","Data":"cac93be348e248204f51c8f1cc8566be80d8ff2432c36b3a8cc3a2d5027aa8f3"} Sep 30 14:43:34 crc kubenswrapper[4799]: I0930 14:43:34.811303 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerStarted","Data":"a5ce25c11cffd6863e280f88bb72199efa4137f1474c52c9e6a0418227bee11d"} Sep 30 14:43:35 crc kubenswrapper[4799]: I0930 14:43:35.342602 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:43:35 crc kubenswrapper[4799]: I0930 14:43:35.343551 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:43:35 crc kubenswrapper[4799]: I0930 14:43:35.395336 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:35 crc kubenswrapper[4799]: I0930 14:43:35.824484 4799 generic.go:334] "Generic (PLEG): container finished" podID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerID="75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b" exitCode=0 Sep 30 14:43:35 crc kubenswrapper[4799]: I0930 14:43:35.824583 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s72t7" event={"ID":"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f","Type":"ContainerDied","Data":"75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b"} Sep 30 14:43:36 crc kubenswrapper[4799]: I0930 14:43:36.842359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s72t7" event={"ID":"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f","Type":"ContainerStarted","Data":"7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57"} Sep 30 14:43:36 crc kubenswrapper[4799]: I0930 14:43:36.849160 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerStarted","Data":"2198559e9d358e59ab9bb81066850933d7320e70e0c1dffa113fc14e7b23236f"} Sep 30 14:43:36 crc kubenswrapper[4799]: I0930 14:43:36.849502 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:43:36 crc kubenswrapper[4799]: I0930 14:43:36.874290 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s72t7" podStartSLOduration=4.058754588 podStartE2EDuration="8.874257495s" podCreationTimestamp="2025-09-30 14:43:28 +0000 UTC" firstStartedPulling="2025-09-30 14:43:31.575257304 +0000 UTC m=+1433.658857741" lastFinishedPulling="2025-09-30 14:43:36.390760211 +0000 UTC m=+1438.474360648" observedRunningTime="2025-09-30 14:43:36.866598591 +0000 UTC m=+1438.950199048" watchObservedRunningTime="2025-09-30 14:43:36.874257495 +0000 UTC m=+1438.957857922" Sep 30 14:43:36 crc kubenswrapper[4799]: I0930 14:43:36.892675 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.310309062 podStartE2EDuration="7.892630802s" podCreationTimestamp="2025-09-30 14:43:29 +0000 UTC" firstStartedPulling="2025-09-30 14:43:31.006129406 +0000 UTC m=+1433.089729843" lastFinishedPulling="2025-09-30 14:43:35.588451166 +0000 UTC m=+1437.672051583" observedRunningTime="2025-09-30 14:43:36.892350724 +0000 UTC m=+1438.975951181" watchObservedRunningTime="2025-09-30 14:43:36.892630802 +0000 UTC m=+1438.976231239" Sep 30 14:43:37 crc kubenswrapper[4799]: I0930 14:43:37.529432 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:43:37 crc kubenswrapper[4799]: I0930 14:43:37.530633 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:43:37 crc kubenswrapper[4799]: I0930 14:43:37.683006 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:43:37 crc kubenswrapper[4799]: I0930 14:43:37.683423 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:43:37 crc kubenswrapper[4799]: I0930 14:43:37.684474 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"7630e75f64223c31df72c9c0ad92ebccc462f64abfa85db6dbbf91edb9c7433c"} pod="openstack/horizon-7f7c888686-pwv2q" containerMessage="Container horizon failed startup probe, will be restarted" Sep 30 14:43:37 crc kubenswrapper[4799]: I0930 14:43:37.684517 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" containerID="cri-o://7630e75f64223c31df72c9c0ad92ebccc462f64abfa85db6dbbf91edb9c7433c" gracePeriod=30 Sep 30 14:43:39 crc kubenswrapper[4799]: I0930 14:43:39.116205 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:39 crc kubenswrapper[4799]: I0930 14:43:39.118089 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:40 crc kubenswrapper[4799]: I0930 14:43:40.172279 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-s72t7" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="registry-server" probeResult="failure" output=< Sep 30 14:43:40 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 14:43:40 crc kubenswrapper[4799]: > Sep 30 14:43:40 crc kubenswrapper[4799]: I0930 14:43:40.343184 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 14:43:40 crc kubenswrapper[4799]: I0930 14:43:40.343604 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 14:43:40 crc kubenswrapper[4799]: I0930 14:43:40.395234 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:40 crc kubenswrapper[4799]: I0930 14:43:40.430428 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:40 crc kubenswrapper[4799]: I0930 14:43:40.830737 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x87rk" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="registry-server" probeResult="failure" output=< Sep 30 14:43:40 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 14:43:40 crc kubenswrapper[4799]: > Sep 30 14:43:40 crc kubenswrapper[4799]: I0930 14:43:40.923930 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.203563 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-8xfj5"] Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.205333 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.213522 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.213876 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.237605 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8xfj5"] Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.239263 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmsws\" (UniqueName: \"kubernetes.io/projected/e0231f24-294e-400b-99ac-2bc5c5e97e25-kube-api-access-pmsws\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.239475 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-config-data\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.239560 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.239638 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-scripts\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.343002 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmsws\" (UniqueName: \"kubernetes.io/projected/e0231f24-294e-400b-99ac-2bc5c5e97e25-kube-api-access-pmsws\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.343155 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-config-data\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.343224 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.343281 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-scripts\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.351603 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.352219 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-config-data\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.358943 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.359441 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.365068 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-scripts\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.385285 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmsws\" (UniqueName: \"kubernetes.io/projected/e0231f24-294e-400b-99ac-2bc5c5e97e25-kube-api-access-pmsws\") pod \"nova-cell1-cell-mapping-8xfj5\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.533975 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.645315 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.645915 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.646434 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.707187 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.931217 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:43:41 crc kubenswrapper[4799]: I0930 14:43:41.952736 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.242043 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-xwzw6"] Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.244635 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.396765 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-xwzw6"] Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.404826 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.406681 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-config\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.406943 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.407091 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.407140 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbldf\" (UniqueName: \"kubernetes.io/projected/9f78baee-06bd-4654-bc64-c9919f4dd420-kube-api-access-hbldf\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.407186 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.444215 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8xfj5"] Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.516475 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.516618 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.516734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbldf\" (UniqueName: \"kubernetes.io/projected/9f78baee-06bd-4654-bc64-c9919f4dd420-kube-api-access-hbldf\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.516811 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.516935 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.517016 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-config\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.517870 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-config\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.524958 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.526106 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.526664 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.527209 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.544918 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbldf\" (UniqueName: \"kubernetes.io/projected/9f78baee-06bd-4654-bc64-c9919f4dd420-kube-api-access-hbldf\") pod \"dnsmasq-dns-59cf4bdb65-xwzw6\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.699233 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:42 crc kubenswrapper[4799]: I0930 14:43:42.948418 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8xfj5" event={"ID":"e0231f24-294e-400b-99ac-2bc5c5e97e25","Type":"ContainerStarted","Data":"5ad62224925aa613bbfc92eb1b2290d54b36a71c9b4b5b6dacc8cab76addb48d"} Sep 30 14:43:43 crc kubenswrapper[4799]: I0930 14:43:43.503626 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-xwzw6"] Sep 30 14:43:43 crc kubenswrapper[4799]: I0930 14:43:43.962700 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8xfj5" event={"ID":"e0231f24-294e-400b-99ac-2bc5c5e97e25","Type":"ContainerStarted","Data":"7942a311c59954911c37813fa7db634c655fb3fae8a1f3706f31a3c67322561b"} Sep 30 14:43:43 crc kubenswrapper[4799]: I0930 14:43:43.970293 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" event={"ID":"9f78baee-06bd-4654-bc64-c9919f4dd420","Type":"ContainerStarted","Data":"d2c142405e8ef5a79c0b5339069481931d2ae923d5541b8088c2980730345262"} Sep 30 14:43:43 crc kubenswrapper[4799]: I0930 14:43:43.996362 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-8xfj5" podStartSLOduration=2.9963288009999998 podStartE2EDuration="2.996328801s" podCreationTimestamp="2025-09-30 14:43:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:43:43.992094127 +0000 UTC m=+1446.075694574" watchObservedRunningTime="2025-09-30 14:43:43.996328801 +0000 UTC m=+1446.079929238" Sep 30 14:43:44 crc kubenswrapper[4799]: I0930 14:43:44.986481 4799 generic.go:334] "Generic (PLEG): container finished" podID="9f78baee-06bd-4654-bc64-c9919f4dd420" containerID="2826ad43b3d354df327ca59f205e8df8bc37f64b5ee8ea1c9f31ab4c93d8cc65" exitCode=0 Sep 30 14:43:44 crc kubenswrapper[4799]: I0930 14:43:44.989054 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" event={"ID":"9f78baee-06bd-4654-bc64-c9919f4dd420","Type":"ContainerDied","Data":"2826ad43b3d354df327ca59f205e8df8bc37f64b5ee8ea1c9f31ab4c93d8cc65"} Sep 30 14:43:46 crc kubenswrapper[4799]: I0930 14:43:46.002599 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" event={"ID":"9f78baee-06bd-4654-bc64-c9919f4dd420","Type":"ContainerStarted","Data":"8b70e418509f54fc6828cf833c4b7035756b4d2c3f7fde7183a6a3c67a912968"} Sep 30 14:43:46 crc kubenswrapper[4799]: I0930 14:43:46.037913 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:46 crc kubenswrapper[4799]: I0930 14:43:46.097463 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" podStartSLOduration=4.097422413 podStartE2EDuration="4.097422413s" podCreationTimestamp="2025-09-30 14:43:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:43:46.0836099 +0000 UTC m=+1448.167210347" watchObservedRunningTime="2025-09-30 14:43:46.097422413 +0000 UTC m=+1448.181022840" Sep 30 14:43:46 crc kubenswrapper[4799]: I0930 14:43:46.652908 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:46 crc kubenswrapper[4799]: I0930 14:43:46.653266 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-log" containerID="cri-o://d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4" gracePeriod=30 Sep 30 14:43:46 crc kubenswrapper[4799]: I0930 14:43:46.653386 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-api" containerID="cri-o://144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c" gracePeriod=30 Sep 30 14:43:47 crc kubenswrapper[4799]: I0930 14:43:47.029203 4799 generic.go:334] "Generic (PLEG): container finished" podID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerID="d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4" exitCode=143 Sep 30 14:43:47 crc kubenswrapper[4799]: I0930 14:43:47.029274 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8","Type":"ContainerDied","Data":"d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4"} Sep 30 14:43:47 crc kubenswrapper[4799]: I0930 14:43:47.537948 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:43:49 crc kubenswrapper[4799]: I0930 14:43:49.197927 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:49 crc kubenswrapper[4799]: I0930 14:43:49.198834 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="ceilometer-central-agent" containerID="cri-o://c46f8ed55b6f42fce896e3a0194f62a1a70b9de0090d8d0e77cee93983e2bdfd" gracePeriod=30 Sep 30 14:43:49 crc kubenswrapper[4799]: I0930 14:43:49.199089 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="sg-core" containerID="cri-o://a5ce25c11cffd6863e280f88bb72199efa4137f1474c52c9e6a0418227bee11d" gracePeriod=30 Sep 30 14:43:49 crc kubenswrapper[4799]: I0930 14:43:49.199330 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="proxy-httpd" containerID="cri-o://2198559e9d358e59ab9bb81066850933d7320e70e0c1dffa113fc14e7b23236f" gracePeriod=30 Sep 30 14:43:49 crc kubenswrapper[4799]: I0930 14:43:49.199354 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="ceilometer-notification-agent" containerID="cri-o://cac93be348e248204f51c8f1cc8566be80d8ff2432c36b3a8cc3a2d5027aa8f3" gracePeriod=30 Sep 30 14:43:49 crc kubenswrapper[4799]: I0930 14:43:49.242449 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.200:3000/\": read tcp 10.217.0.2:35548->10.217.0.200:3000: read: connection reset by peer" Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.076973 4799 generic.go:334] "Generic (PLEG): container finished" podID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerID="2198559e9d358e59ab9bb81066850933d7320e70e0c1dffa113fc14e7b23236f" exitCode=0 Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.077354 4799 generic.go:334] "Generic (PLEG): container finished" podID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerID="a5ce25c11cffd6863e280f88bb72199efa4137f1474c52c9e6a0418227bee11d" exitCode=2 Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.077368 4799 generic.go:334] "Generic (PLEG): container finished" podID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerID="cac93be348e248204f51c8f1cc8566be80d8ff2432c36b3a8cc3a2d5027aa8f3" exitCode=0 Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.077046 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerDied","Data":"2198559e9d358e59ab9bb81066850933d7320e70e0c1dffa113fc14e7b23236f"} Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.077412 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerDied","Data":"a5ce25c11cffd6863e280f88bb72199efa4137f1474c52c9e6a0418227bee11d"} Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.077425 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerDied","Data":"cac93be348e248204f51c8f1cc8566be80d8ff2432c36b3a8cc3a2d5027aa8f3"} Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.193943 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-s72t7" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="registry-server" probeResult="failure" output=< Sep 30 14:43:50 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 14:43:50 crc kubenswrapper[4799]: > Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.378216 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.381946 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.405329 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.849176 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x87rk" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="registry-server" probeResult="failure" output=< Sep 30 14:43:50 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 14:43:50 crc kubenswrapper[4799]: > Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.890449 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.982073 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-config-data\") pod \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.982218 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z49pk\" (UniqueName: \"kubernetes.io/projected/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-kube-api-access-z49pk\") pod \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.982380 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-logs\") pod \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.982407 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-combined-ca-bundle\") pod \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\" (UID: \"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8\") " Sep 30 14:43:50 crc kubenswrapper[4799]: I0930 14:43:50.984368 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-logs" (OuterVolumeSpecName: "logs") pod "34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" (UID: "34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.000112 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-kube-api-access-z49pk" (OuterVolumeSpecName: "kube-api-access-z49pk") pod "34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" (UID: "34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8"). InnerVolumeSpecName "kube-api-access-z49pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.056926 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-config-data" (OuterVolumeSpecName: "config-data") pod "34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" (UID: "34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.067596 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" (UID: "34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.122524 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.122555 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z49pk\" (UniqueName: \"kubernetes.io/projected/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-kube-api-access-z49pk\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.122568 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.122577 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.168568 4799 generic.go:334] "Generic (PLEG): container finished" podID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerID="144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c" exitCode=0 Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.169848 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.170329 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8","Type":"ContainerDied","Data":"144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c"} Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.170390 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8","Type":"ContainerDied","Data":"24b9134b5558104dd356b8966b050fc930230722e072f62034a03cde42eed265"} Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.170411 4799 scope.go:117] "RemoveContainer" containerID="144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.183478 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.273743 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.290818 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.298901 4799 scope.go:117] "RemoveContainer" containerID="d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.332615 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:51 crc kubenswrapper[4799]: E0930 14:43:51.333572 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-api" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.333595 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-api" Sep 30 14:43:51 crc kubenswrapper[4799]: E0930 14:43:51.333618 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-log" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.333625 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-log" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.333830 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-api" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.333857 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" containerName="nova-api-log" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.335135 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.343844 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.343935 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.344010 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.354488 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.411021 4799 scope.go:117] "RemoveContainer" containerID="144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c" Sep 30 14:43:51 crc kubenswrapper[4799]: E0930 14:43:51.412620 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c\": container with ID starting with 144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c not found: ID does not exist" containerID="144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.412681 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c"} err="failed to get container status \"144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c\": rpc error: code = NotFound desc = could not find container \"144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c\": container with ID starting with 144170be26fcab5effdb25233a32733a1fb15254b91bc3e2623bc556589e136c not found: ID does not exist" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.412712 4799 scope.go:117] "RemoveContainer" containerID="d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4" Sep 30 14:43:51 crc kubenswrapper[4799]: E0930 14:43:51.413056 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4\": container with ID starting with d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4 not found: ID does not exist" containerID="d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.413079 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4"} err="failed to get container status \"d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4\": rpc error: code = NotFound desc = could not find container \"d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4\": container with ID starting with d184f439da8a38f98c09f0afeb9b6ff9dc8c5d3c37ad342d9a14e197efa577b4 not found: ID does not exist" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.434613 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.434729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-logs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.434921 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.435084 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.435394 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-config-data\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.435448 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff4qh\" (UniqueName: \"kubernetes.io/projected/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-kube-api-access-ff4qh\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.537721 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.537846 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-logs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.537883 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.537926 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.538019 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-config-data\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.538046 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff4qh\" (UniqueName: \"kubernetes.io/projected/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-kube-api-access-ff4qh\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.538507 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-logs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.543965 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.546584 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-config-data\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.550217 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.550343 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.560181 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff4qh\" (UniqueName: \"kubernetes.io/projected/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-kube-api-access-ff4qh\") pod \"nova-api-0\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " pod="openstack/nova-api-0" Sep 30 14:43:51 crc kubenswrapper[4799]: I0930 14:43:51.714350 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.264215 4799 generic.go:334] "Generic (PLEG): container finished" podID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerID="c46f8ed55b6f42fce896e3a0194f62a1a70b9de0090d8d0e77cee93983e2bdfd" exitCode=0 Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.264752 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerDied","Data":"c46f8ed55b6f42fce896e3a0194f62a1a70b9de0090d8d0e77cee93983e2bdfd"} Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.464286 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.486213 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:43:52 crc kubenswrapper[4799]: W0930 14:43:52.491448 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod590a9114_0e5d_41d7_ad2a_8c4cdd4fd8ee.slice/crio-c3cfe43d4d782c7bf57f3bc5c9793a67079564caa9239ba1c03c751a0f976d7e WatchSource:0}: Error finding container c3cfe43d4d782c7bf57f3bc5c9793a67079564caa9239ba1c03c751a0f976d7e: Status 404 returned error can't find the container with id c3cfe43d4d782c7bf57f3bc5c9793a67079564caa9239ba1c03c751a0f976d7e Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.544544 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8" path="/var/lib/kubelet/pods/34ed5a01-f6ba-4e1e-b13b-5070d7f68ec8/volumes" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.607465 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-config-data\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.608102 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-ceilometer-tls-certs\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.608162 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-log-httpd\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.608295 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g2rw\" (UniqueName: \"kubernetes.io/projected/5035aa23-c338-4ae2-8de9-94ac166052c9-kube-api-access-2g2rw\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.608377 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-run-httpd\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.608401 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-sg-core-conf-yaml\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.608445 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-combined-ca-bundle\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.608507 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-scripts\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.610577 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.611200 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.626094 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-scripts" (OuterVolumeSpecName: "scripts") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.631130 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5035aa23-c338-4ae2-8de9-94ac166052c9-kube-api-access-2g2rw" (OuterVolumeSpecName: "kube-api-access-2g2rw") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "kube-api-access-2g2rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.712246 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.712290 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g2rw\" (UniqueName: \"kubernetes.io/projected/5035aa23-c338-4ae2-8de9-94ac166052c9-kube-api-access-2g2rw\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.712306 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5035aa23-c338-4ae2-8de9-94ac166052c9-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.712318 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.713260 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.818113 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.831426 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-ceilometer-tls-certs\") pod \"5035aa23-c338-4ae2-8de9-94ac166052c9\" (UID: \"5035aa23-c338-4ae2-8de9-94ac166052c9\") " Sep 30 14:43:52 crc kubenswrapper[4799]: W0930 14:43:52.833292 4799 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5035aa23-c338-4ae2-8de9-94ac166052c9/volumes/kubernetes.io~secret/ceilometer-tls-certs Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.833324 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.877818 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.909518 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xv9g7"] Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.919577 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" podUID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerName="dnsmasq-dns" containerID="cri-o://b7b735aabdc6ea4e401e3ef0ca463579b572fae1e090a735c4aa413a219bbe81" gracePeriod=10 Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.936425 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.936477 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:52 crc kubenswrapper[4799]: I0930 14:43:52.989518 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.038251 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.111815 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-config-data" (OuterVolumeSpecName: "config-data") pod "5035aa23-c338-4ae2-8de9-94ac166052c9" (UID: "5035aa23-c338-4ae2-8de9-94ac166052c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.140842 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5035aa23-c338-4ae2-8de9-94ac166052c9-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.290296 4799 generic.go:334] "Generic (PLEG): container finished" podID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerID="b7b735aabdc6ea4e401e3ef0ca463579b572fae1e090a735c4aa413a219bbe81" exitCode=0 Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.290390 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" event={"ID":"21856f94-450c-4c1a-9cf9-7697d7d22a44","Type":"ContainerDied","Data":"b7b735aabdc6ea4e401e3ef0ca463579b572fae1e090a735c4aa413a219bbe81"} Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.316357 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5035aa23-c338-4ae2-8de9-94ac166052c9","Type":"ContainerDied","Data":"fc54f687230fc12b273b42d1f7a4f8486a65f34d54d3735917b86b4436fe689d"} Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.316417 4799 scope.go:117] "RemoveContainer" containerID="2198559e9d358e59ab9bb81066850933d7320e70e0c1dffa113fc14e7b23236f" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.316569 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.336199 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee","Type":"ContainerStarted","Data":"c3cfe43d4d782c7bf57f3bc5c9793a67079564caa9239ba1c03c751a0f976d7e"} Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.416973 4799 scope.go:117] "RemoveContainer" containerID="a5ce25c11cffd6863e280f88bb72199efa4137f1474c52c9e6a0418227bee11d" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.419684 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.459367 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.466798 4799 scope.go:117] "RemoveContainer" containerID="cac93be348e248204f51c8f1cc8566be80d8ff2432c36b3a8cc3a2d5027aa8f3" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.522491 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:53 crc kubenswrapper[4799]: E0930 14:43:53.523634 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="ceilometer-central-agent" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.523685 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="ceilometer-central-agent" Sep 30 14:43:53 crc kubenswrapper[4799]: E0930 14:43:53.523711 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="ceilometer-notification-agent" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.523720 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="ceilometer-notification-agent" Sep 30 14:43:53 crc kubenswrapper[4799]: E0930 14:43:53.523737 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="proxy-httpd" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.523747 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="proxy-httpd" Sep 30 14:43:53 crc kubenswrapper[4799]: E0930 14:43:53.524962 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="sg-core" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.524980 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="sg-core" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.525690 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="proxy-httpd" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.525988 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="sg-core" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.526181 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="ceilometer-central-agent" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.526330 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" containerName="ceilometer-notification-agent" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.539568 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.546087 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.551233 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.551596 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.551638 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.580868 4799 scope.go:117] "RemoveContainer" containerID="c46f8ed55b6f42fce896e3a0194f62a1a70b9de0090d8d0e77cee93983e2bdfd" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.665135 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6spnh\" (UniqueName: \"kubernetes.io/projected/77b5c423-17cd-416f-91cf-56aa87947c5c-kube-api-access-6spnh\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.665244 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.665281 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.665325 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-run-httpd\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.665374 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-config-data\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.665437 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.665459 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-scripts\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.665487 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-log-httpd\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.767679 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.768065 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.768136 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-run-httpd\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.768228 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-config-data\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.768276 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.768301 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-scripts\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.768331 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-log-httpd\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.768383 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6spnh\" (UniqueName: \"kubernetes.io/projected/77b5c423-17cd-416f-91cf-56aa87947c5c-kube-api-access-6spnh\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.775160 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-run-httpd\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.775183 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-log-httpd\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.781211 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.781636 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-config-data\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.781968 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.786212 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-scripts\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.788395 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.792838 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6spnh\" (UniqueName: \"kubernetes.io/projected/77b5c423-17cd-416f-91cf-56aa87947c5c-kube-api-access-6spnh\") pod \"ceilometer-0\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.910981 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.945186 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.973296 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-svc\") pod \"21856f94-450c-4c1a-9cf9-7697d7d22a44\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.973366 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-sb\") pod \"21856f94-450c-4c1a-9cf9-7697d7d22a44\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.973436 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm85v\" (UniqueName: \"kubernetes.io/projected/21856f94-450c-4c1a-9cf9-7697d7d22a44-kube-api-access-cm85v\") pod \"21856f94-450c-4c1a-9cf9-7697d7d22a44\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.973475 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-config\") pod \"21856f94-450c-4c1a-9cf9-7697d7d22a44\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.973696 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-swift-storage-0\") pod \"21856f94-450c-4c1a-9cf9-7697d7d22a44\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " Sep 30 14:43:53 crc kubenswrapper[4799]: I0930 14:43:53.973739 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-nb\") pod \"21856f94-450c-4c1a-9cf9-7697d7d22a44\" (UID: \"21856f94-450c-4c1a-9cf9-7697d7d22a44\") " Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.009800 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21856f94-450c-4c1a-9cf9-7697d7d22a44-kube-api-access-cm85v" (OuterVolumeSpecName: "kube-api-access-cm85v") pod "21856f94-450c-4c1a-9cf9-7697d7d22a44" (UID: "21856f94-450c-4c1a-9cf9-7697d7d22a44"). InnerVolumeSpecName "kube-api-access-cm85v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.078854 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm85v\" (UniqueName: \"kubernetes.io/projected/21856f94-450c-4c1a-9cf9-7697d7d22a44-kube-api-access-cm85v\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.174134 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21856f94-450c-4c1a-9cf9-7697d7d22a44" (UID: "21856f94-450c-4c1a-9cf9-7697d7d22a44"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.183936 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.187786 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-config" (OuterVolumeSpecName: "config") pod "21856f94-450c-4c1a-9cf9-7697d7d22a44" (UID: "21856f94-450c-4c1a-9cf9-7697d7d22a44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.194459 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "21856f94-450c-4c1a-9cf9-7697d7d22a44" (UID: "21856f94-450c-4c1a-9cf9-7697d7d22a44"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.227767 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21856f94-450c-4c1a-9cf9-7697d7d22a44" (UID: "21856f94-450c-4c1a-9cf9-7697d7d22a44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.240228 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21856f94-450c-4c1a-9cf9-7697d7d22a44" (UID: "21856f94-450c-4c1a-9cf9-7697d7d22a44"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.285520 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.285578 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.285592 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.285607 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21856f94-450c-4c1a-9cf9-7697d7d22a44-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.387855 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" event={"ID":"21856f94-450c-4c1a-9cf9-7697d7d22a44","Type":"ContainerDied","Data":"9dac06f32e2c1ce83297b3a5f21d0f04ef0d9d27a99eaeb0c02b2d76eeca628c"} Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.387930 4799 scope.go:117] "RemoveContainer" containerID="b7b735aabdc6ea4e401e3ef0ca463579b572fae1e090a735c4aa413a219bbe81" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.388091 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.424450 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee","Type":"ContainerStarted","Data":"e02238ae5c743030c91919d0ab64563a90715b6438cfaae5488a04b00429122e"} Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.424514 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee","Type":"ContainerStarted","Data":"a70a25564fdb3b67d32edf0d75840eca63f5349a68944a79765020c9e55d9e74"} Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.474818 4799 scope.go:117] "RemoveContainer" containerID="4538be714431660e59b8ffe63642f74a14da10ad42592f9022ea6e1f60e5433e" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.499604 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.499574232 podStartE2EDuration="3.499574232s" podCreationTimestamp="2025-09-30 14:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:43:54.474273883 +0000 UTC m=+1456.557874330" watchObservedRunningTime="2025-09-30 14:43:54.499574232 +0000 UTC m=+1456.583174659" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.555506 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5035aa23-c338-4ae2-8de9-94ac166052c9" path="/var/lib/kubelet/pods/5035aa23-c338-4ae2-8de9-94ac166052c9/volumes" Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.559559 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xv9g7"] Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.559615 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-xv9g7"] Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.631678 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:54 crc kubenswrapper[4799]: I0930 14:43:54.919028 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:43:55 crc kubenswrapper[4799]: I0930 14:43:55.438499 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerStarted","Data":"331947af2a2d948861686f78dac43e8c8292bf5feb4677d3c4107e98c970c881"} Sep 30 14:43:56 crc kubenswrapper[4799]: I0930 14:43:56.562770 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21856f94-450c-4c1a-9cf9-7697d7d22a44" path="/var/lib/kubelet/pods/21856f94-450c-4c1a-9cf9-7697d7d22a44/volumes" Sep 30 14:43:56 crc kubenswrapper[4799]: I0930 14:43:56.566951 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerStarted","Data":"bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736"} Sep 30 14:43:57 crc kubenswrapper[4799]: I0930 14:43:57.529220 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:43:57 crc kubenswrapper[4799]: I0930 14:43:57.538007 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerStarted","Data":"84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73"} Sep 30 14:43:58 crc kubenswrapper[4799]: I0930 14:43:58.461008 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-845d6d6f59-xv9g7" podUID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.191:5353: i/o timeout" Sep 30 14:43:58 crc kubenswrapper[4799]: I0930 14:43:58.551678 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerStarted","Data":"e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c"} Sep 30 14:43:58 crc kubenswrapper[4799]: I0930 14:43:58.554724 4799 generic.go:334] "Generic (PLEG): container finished" podID="e0231f24-294e-400b-99ac-2bc5c5e97e25" containerID="7942a311c59954911c37813fa7db634c655fb3fae8a1f3706f31a3c67322561b" exitCode=0 Sep 30 14:43:58 crc kubenswrapper[4799]: I0930 14:43:58.554779 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8xfj5" event={"ID":"e0231f24-294e-400b-99ac-2bc5c5e97e25","Type":"ContainerDied","Data":"7942a311c59954911c37813fa7db634c655fb3fae8a1f3706f31a3c67322561b"} Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.183216 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.250695 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.587246 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerStarted","Data":"2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402"} Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.588883 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="ceilometer-central-agent" containerID="cri-o://bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736" gracePeriod=30 Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.589449 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="proxy-httpd" containerID="cri-o://2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402" gracePeriod=30 Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.589521 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="sg-core" containerID="cri-o://e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c" gracePeriod=30 Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.589583 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="ceilometer-notification-agent" containerID="cri-o://84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73" gracePeriod=30 Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.637902 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.519013505 podStartE2EDuration="6.637871404s" podCreationTimestamp="2025-09-30 14:43:53 +0000 UTC" firstStartedPulling="2025-09-30 14:43:54.681879482 +0000 UTC m=+1456.765479909" lastFinishedPulling="2025-09-30 14:43:58.800737381 +0000 UTC m=+1460.884337808" observedRunningTime="2025-09-30 14:43:59.636751621 +0000 UTC m=+1461.720352058" watchObservedRunningTime="2025-09-30 14:43:59.637871404 +0000 UTC m=+1461.721471831" Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.852739 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:43:59 crc kubenswrapper[4799]: I0930 14:43:59.932259 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s72t7"] Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.005731 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.290784 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.385275 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-config-data\") pod \"e0231f24-294e-400b-99ac-2bc5c5e97e25\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.385672 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-combined-ca-bundle\") pod \"e0231f24-294e-400b-99ac-2bc5c5e97e25\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.385739 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmsws\" (UniqueName: \"kubernetes.io/projected/e0231f24-294e-400b-99ac-2bc5c5e97e25-kube-api-access-pmsws\") pod \"e0231f24-294e-400b-99ac-2bc5c5e97e25\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.385958 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-scripts\") pod \"e0231f24-294e-400b-99ac-2bc5c5e97e25\" (UID: \"e0231f24-294e-400b-99ac-2bc5c5e97e25\") " Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.392621 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-scripts" (OuterVolumeSpecName: "scripts") pod "e0231f24-294e-400b-99ac-2bc5c5e97e25" (UID: "e0231f24-294e-400b-99ac-2bc5c5e97e25"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.440599 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0231f24-294e-400b-99ac-2bc5c5e97e25" (UID: "e0231f24-294e-400b-99ac-2bc5c5e97e25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.442858 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0231f24-294e-400b-99ac-2bc5c5e97e25-kube-api-access-pmsws" (OuterVolumeSpecName: "kube-api-access-pmsws") pod "e0231f24-294e-400b-99ac-2bc5c5e97e25" (UID: "e0231f24-294e-400b-99ac-2bc5c5e97e25"). InnerVolumeSpecName "kube-api-access-pmsws". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.487725 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-config-data" (OuterVolumeSpecName: "config-data") pod "e0231f24-294e-400b-99ac-2bc5c5e97e25" (UID: "e0231f24-294e-400b-99ac-2bc5c5e97e25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.488483 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.488587 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.488700 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmsws\" (UniqueName: \"kubernetes.io/projected/e0231f24-294e-400b-99ac-2bc5c5e97e25-kube-api-access-pmsws\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.488769 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0231f24-294e-400b-99ac-2bc5c5e97e25-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.614630 4799 generic.go:334] "Generic (PLEG): container finished" podID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerID="2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402" exitCode=0 Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.614705 4799 generic.go:334] "Generic (PLEG): container finished" podID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerID="e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c" exitCode=2 Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.614776 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerDied","Data":"2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402"} Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.614816 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerDied","Data":"e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c"} Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.620205 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s72t7" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="registry-server" containerID="cri-o://7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57" gracePeriod=2 Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.620642 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8xfj5" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.622326 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8xfj5" event={"ID":"e0231f24-294e-400b-99ac-2bc5c5e97e25","Type":"ContainerDied","Data":"5ad62224925aa613bbfc92eb1b2290d54b36a71c9b4b5b6dacc8cab76addb48d"} Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.622378 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ad62224925aa613bbfc92eb1b2290d54b36a71c9b4b5b6dacc8cab76addb48d" Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.785290 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.785950 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerName="nova-api-log" containerID="cri-o://a70a25564fdb3b67d32edf0d75840eca63f5349a68944a79765020c9e55d9e74" gracePeriod=30 Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.786143 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerName="nova-api-api" containerID="cri-o://e02238ae5c743030c91919d0ab64563a90715b6438cfaae5488a04b00429122e" gracePeriod=30 Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.805617 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.805918 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d7a4ded6-46db-4145-a518-f9b3dee3843e" containerName="nova-scheduler-scheduler" containerID="cri-o://21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04" gracePeriod=30 Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.843760 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.844105 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-log" containerID="cri-o://d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7" gracePeriod=30 Sep 30 14:44:00 crc kubenswrapper[4799]: I0930 14:44:00.844756 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-metadata" containerID="cri-o://fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44" gracePeriod=30 Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.342728 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.421384 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q9m2\" (UniqueName: \"kubernetes.io/projected/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-kube-api-access-6q9m2\") pod \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.421506 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-utilities\") pod \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.421544 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-catalog-content\") pod \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\" (UID: \"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.423036 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-utilities" (OuterVolumeSpecName: "utilities") pod "2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" (UID: "2ec75a09-9dc0-4aee-a6fe-c7748d489b5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.442782 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-kube-api-access-6q9m2" (OuterVolumeSpecName: "kube-api-access-6q9m2") pod "2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" (UID: "2ec75a09-9dc0-4aee-a6fe-c7748d489b5f"). InnerVolumeSpecName "kube-api-access-6q9m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.494320 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" (UID: "2ec75a09-9dc0-4aee-a6fe-c7748d489b5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.525957 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.526018 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.526032 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q9m2\" (UniqueName: \"kubernetes.io/projected/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f-kube-api-access-6q9m2\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.634629 4799 generic.go:334] "Generic (PLEG): container finished" podID="ce901e13-dfa1-4b79-a584-4903008803d0" containerID="d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7" exitCode=143 Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.634741 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce901e13-dfa1-4b79-a584-4903008803d0","Type":"ContainerDied","Data":"d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7"} Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.642586 4799 generic.go:334] "Generic (PLEG): container finished" podID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerID="e02238ae5c743030c91919d0ab64563a90715b6438cfaae5488a04b00429122e" exitCode=0 Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.642643 4799 generic.go:334] "Generic (PLEG): container finished" podID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerID="a70a25564fdb3b67d32edf0d75840eca63f5349a68944a79765020c9e55d9e74" exitCode=143 Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.642741 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee","Type":"ContainerDied","Data":"e02238ae5c743030c91919d0ab64563a90715b6438cfaae5488a04b00429122e"} Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.642781 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee","Type":"ContainerDied","Data":"a70a25564fdb3b67d32edf0d75840eca63f5349a68944a79765020c9e55d9e74"} Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.659039 4799 generic.go:334] "Generic (PLEG): container finished" podID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerID="84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73" exitCode=0 Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.659310 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerDied","Data":"84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73"} Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.667314 4799 generic.go:334] "Generic (PLEG): container finished" podID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerID="7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57" exitCode=0 Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.667401 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s72t7" event={"ID":"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f","Type":"ContainerDied","Data":"7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57"} Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.667465 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s72t7" event={"ID":"2ec75a09-9dc0-4aee-a6fe-c7748d489b5f","Type":"ContainerDied","Data":"39c99460ce9195bd3d8a63eef52e21b38f6334e37370ea8d91f9f0cc7c8793c0"} Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.667495 4799 scope.go:117] "RemoveContainer" containerID="7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.667997 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s72t7" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.705253 4799 scope.go:117] "RemoveContainer" containerID="75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.736402 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s72t7"] Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.748984 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s72t7"] Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.751966 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.764299 4799 scope.go:117] "RemoveContainer" containerID="7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.808385 4799 scope.go:117] "RemoveContainer" containerID="7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57" Sep 30 14:44:01 crc kubenswrapper[4799]: E0930 14:44:01.808936 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57\": container with ID starting with 7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57 not found: ID does not exist" containerID="7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.808966 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57"} err="failed to get container status \"7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57\": rpc error: code = NotFound desc = could not find container \"7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57\": container with ID starting with 7881fdc691028e8b979d19ab49e6488cdcb7ff2c7737f86c274c756710923c57 not found: ID does not exist" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.808994 4799 scope.go:117] "RemoveContainer" containerID="75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b" Sep 30 14:44:01 crc kubenswrapper[4799]: E0930 14:44:01.809185 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b\": container with ID starting with 75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b not found: ID does not exist" containerID="75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.809205 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b"} err="failed to get container status \"75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b\": rpc error: code = NotFound desc = could not find container \"75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b\": container with ID starting with 75be7d64e1f9dc49eec946d3b8cbd7a488df75540225cd4b7ef5f6bb0e19a84b not found: ID does not exist" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.809237 4799 scope.go:117] "RemoveContainer" containerID="7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4" Sep 30 14:44:01 crc kubenswrapper[4799]: E0930 14:44:01.809424 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4\": container with ID starting with 7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4 not found: ID does not exist" containerID="7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.809442 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4"} err="failed to get container status \"7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4\": rpc error: code = NotFound desc = could not find container \"7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4\": container with ID starting with 7897e3517975aa9a998d304eaf94fcaf9da9f5b57dd74d5f1feca9e28f4445c4 not found: ID does not exist" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.833515 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-logs\") pod \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.833612 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-config-data\") pod \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.833637 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-internal-tls-certs\") pod \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.833689 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff4qh\" (UniqueName: \"kubernetes.io/projected/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-kube-api-access-ff4qh\") pod \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.833779 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-combined-ca-bundle\") pod \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.833837 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-public-tls-certs\") pod \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\" (UID: \"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee\") " Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.834349 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-logs" (OuterVolumeSpecName: "logs") pod "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" (UID: "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.838331 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.856953 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-kube-api-access-ff4qh" (OuterVolumeSpecName: "kube-api-access-ff4qh") pod "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" (UID: "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee"). InnerVolumeSpecName "kube-api-access-ff4qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.877949 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-config-data" (OuterVolumeSpecName: "config-data") pod "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" (UID: "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.879811 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" (UID: "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.940863 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.940945 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff4qh\" (UniqueName: \"kubernetes.io/projected/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-kube-api-access-ff4qh\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.940960 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.943183 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" (UID: "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:01 crc kubenswrapper[4799]: I0930 14:44:01.952156 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" (UID: "590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.043329 4799 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.043385 4799 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee-public-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.150197 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x87rk"] Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.150912 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x87rk" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="registry-server" containerID="cri-o://cb7ef54d9ceb77e889a553ae7f46f32011c58b4d366a7258f2be9bc982883bcd" gracePeriod=2 Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.520393 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" path="/var/lib/kubelet/pods/2ec75a09-9dc0-4aee-a6fe-c7748d489b5f/volumes" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.758046 4799 generic.go:334] "Generic (PLEG): container finished" podID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerID="cb7ef54d9ceb77e889a553ae7f46f32011c58b4d366a7258f2be9bc982883bcd" exitCode=0 Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.758179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x87rk" event={"ID":"8f6380d4-aaf5-41ac-b134-9f46269f2183","Type":"ContainerDied","Data":"cb7ef54d9ceb77e889a553ae7f46f32011c58b4d366a7258f2be9bc982883bcd"} Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.770385 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee","Type":"ContainerDied","Data":"c3cfe43d4d782c7bf57f3bc5c9793a67079564caa9239ba1c03c751a0f976d7e"} Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.770463 4799 scope.go:117] "RemoveContainer" containerID="e02238ae5c743030c91919d0ab64563a90715b6438cfaae5488a04b00429122e" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.770747 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.842267 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.853002 4799 scope.go:117] "RemoveContainer" containerID="a70a25564fdb3b67d32edf0d75840eca63f5349a68944a79765020c9e55d9e74" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.870967 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.889367 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Sep 30 14:44:02 crc kubenswrapper[4799]: E0930 14:44:02.890349 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerName="dnsmasq-dns" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890377 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerName="dnsmasq-dns" Sep 30 14:44:02 crc kubenswrapper[4799]: E0930 14:44:02.890399 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="extract-utilities" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890408 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="extract-utilities" Sep 30 14:44:02 crc kubenswrapper[4799]: E0930 14:44:02.890429 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="extract-content" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890438 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="extract-content" Sep 30 14:44:02 crc kubenswrapper[4799]: E0930 14:44:02.890465 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0231f24-294e-400b-99ac-2bc5c5e97e25" containerName="nova-manage" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890474 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0231f24-294e-400b-99ac-2bc5c5e97e25" containerName="nova-manage" Sep 30 14:44:02 crc kubenswrapper[4799]: E0930 14:44:02.890486 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerName="nova-api-log" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890494 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerName="nova-api-log" Sep 30 14:44:02 crc kubenswrapper[4799]: E0930 14:44:02.890509 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="registry-server" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890518 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="registry-server" Sep 30 14:44:02 crc kubenswrapper[4799]: E0930 14:44:02.890536 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerName="init" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890544 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerName="init" Sep 30 14:44:02 crc kubenswrapper[4799]: E0930 14:44:02.890567 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerName="nova-api-api" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890575 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerName="nova-api-api" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890839 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ec75a09-9dc0-4aee-a6fe-c7748d489b5f" containerName="registry-server" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890864 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0231f24-294e-400b-99ac-2bc5c5e97e25" containerName="nova-manage" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890883 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerName="nova-api-api" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890905 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" containerName="nova-api-log" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.890913 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="21856f94-450c-4c1a-9cf9-7697d7d22a44" containerName="dnsmasq-dns" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.892504 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.900741 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.903524 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.904802 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.905040 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.969878 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-public-tls-certs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.969970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.970003 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28bc13f2-5895-42cf-8ab5-60251eb4d09f-logs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.970031 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-config-data\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.970095 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbks\" (UniqueName: \"kubernetes.io/projected/28bc13f2-5895-42cf-8ab5-60251eb4d09f-kube-api-access-pjbks\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:02 crc kubenswrapper[4799]: I0930 14:44:02.970211 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.050782 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.071791 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-utilities\") pod \"8f6380d4-aaf5-41ac-b134-9f46269f2183\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.072117 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-catalog-content\") pod \"8f6380d4-aaf5-41ac-b134-9f46269f2183\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.072360 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktr7p\" (UniqueName: \"kubernetes.io/projected/8f6380d4-aaf5-41ac-b134-9f46269f2183-kube-api-access-ktr7p\") pod \"8f6380d4-aaf5-41ac-b134-9f46269f2183\" (UID: \"8f6380d4-aaf5-41ac-b134-9f46269f2183\") " Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.073153 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.073401 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-public-tls-certs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.073571 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.073686 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28bc13f2-5895-42cf-8ab5-60251eb4d09f-logs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.073910 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-config-data\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.074092 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbks\" (UniqueName: \"kubernetes.io/projected/28bc13f2-5895-42cf-8ab5-60251eb4d09f-kube-api-access-pjbks\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.073796 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-utilities" (OuterVolumeSpecName: "utilities") pod "8f6380d4-aaf5-41ac-b134-9f46269f2183" (UID: "8f6380d4-aaf5-41ac-b134-9f46269f2183"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.075377 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28bc13f2-5895-42cf-8ab5-60251eb4d09f-logs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.083999 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-public-tls-certs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.097332 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-config-data\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.099869 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f6380d4-aaf5-41ac-b134-9f46269f2183-kube-api-access-ktr7p" (OuterVolumeSpecName: "kube-api-access-ktr7p") pod "8f6380d4-aaf5-41ac-b134-9f46269f2183" (UID: "8f6380d4-aaf5-41ac-b134-9f46269f2183"). InnerVolumeSpecName "kube-api-access-ktr7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.123865 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbks\" (UniqueName: \"kubernetes.io/projected/28bc13f2-5895-42cf-8ab5-60251eb4d09f-kube-api-access-pjbks\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.124185 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.124199 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bc13f2-5895-42cf-8ab5-60251eb4d09f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"28bc13f2-5895-42cf-8ab5-60251eb4d09f\") " pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.176193 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.176240 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktr7p\" (UniqueName: \"kubernetes.io/projected/8f6380d4-aaf5-41ac-b134-9f46269f2183-kube-api-access-ktr7p\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.214622 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f6380d4-aaf5-41ac-b134-9f46269f2183" (UID: "8f6380d4-aaf5-41ac-b134-9f46269f2183"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.280532 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f6380d4-aaf5-41ac-b134-9f46269f2183-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.341927 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.789934 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x87rk" event={"ID":"8f6380d4-aaf5-41ac-b134-9f46269f2183","Type":"ContainerDied","Data":"cedd1bf8ca6758888b0bf2c287ee8122ad8174dc996f2ae43a9efb20e3d99a78"} Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.790208 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x87rk" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.790296 4799 scope.go:117] "RemoveContainer" containerID="cb7ef54d9ceb77e889a553ae7f46f32011c58b4d366a7258f2be9bc982883bcd" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.850584 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x87rk"] Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.852700 4799 scope.go:117] "RemoveContainer" containerID="b11ad949030a57b82b99be46ae919d927b149a5f302688076abb0b63cfdf85ff" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.862498 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x87rk"] Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.888752 4799 scope.go:117] "RemoveContainer" containerID="00f941859362641f79192a151fc04c974ae808d102269a14feb1bcbcdaa0a92a" Sep 30 14:44:03 crc kubenswrapper[4799]: I0930 14:44:03.910984 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Sep 30 14:44:03 crc kubenswrapper[4799]: W0930 14:44:03.918628 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28bc13f2_5895_42cf_8ab5_60251eb4d09f.slice/crio-48e1ff9f4e7ce7f914e09bfe9b1b106047ad4d93bde45fea8ea2fd5a20da3a51 WatchSource:0}: Error finding container 48e1ff9f4e7ce7f914e09bfe9b1b106047ad4d93bde45fea8ea2fd5a20da3a51: Status 404 returned error can't find the container with id 48e1ff9f4e7ce7f914e09bfe9b1b106047ad4d93bde45fea8ea2fd5a20da3a51 Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.548609 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee" path="/var/lib/kubelet/pods/590a9114-0e5d-41d7-ad2a-8c4cdd4fd8ee/volumes" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.550057 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" path="/var/lib/kubelet/pods/8f6380d4-aaf5-41ac-b134-9f46269f2183/volumes" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.775805 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.808294 4799 generic.go:334] "Generic (PLEG): container finished" podID="ce901e13-dfa1-4b79-a584-4903008803d0" containerID="fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44" exitCode=0 Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.808387 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce901e13-dfa1-4b79-a584-4903008803d0","Type":"ContainerDied","Data":"fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44"} Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.808432 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce901e13-dfa1-4b79-a584-4903008803d0","Type":"ContainerDied","Data":"eb27a9bd03bfd7cd7fdda4caec0b0117181b206f4f17304fd413ae0edf44237b"} Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.808463 4799 scope.go:117] "RemoveContainer" containerID="fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.808680 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.812860 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"28bc13f2-5895-42cf-8ab5-60251eb4d09f","Type":"ContainerStarted","Data":"cb60ea72263aff994f14d7452111cbfb07a9241e047ec0455759a3bfbe811687"} Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.812920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"28bc13f2-5895-42cf-8ab5-60251eb4d09f","Type":"ContainerStarted","Data":"b0b5c5d1dafdb28dc2c0eeeb0dfb6ab3d071aa24a48dd744ae6fc2667f2a1e37"} Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.812938 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"28bc13f2-5895-42cf-8ab5-60251eb4d09f","Type":"ContainerStarted","Data":"48e1ff9f4e7ce7f914e09bfe9b1b106047ad4d93bde45fea8ea2fd5a20da3a51"} Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.858266 4799 scope.go:117] "RemoveContainer" containerID="d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.871908 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.871879285 podStartE2EDuration="2.871879285s" podCreationTimestamp="2025-09-30 14:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:44:04.866312372 +0000 UTC m=+1466.949912819" watchObservedRunningTime="2025-09-30 14:44:04.871879285 +0000 UTC m=+1466.955479702" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.895166 4799 scope.go:117] "RemoveContainer" containerID="fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44" Sep 30 14:44:04 crc kubenswrapper[4799]: E0930 14:44:04.895837 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44\": container with ID starting with fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44 not found: ID does not exist" containerID="fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.895906 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44"} err="failed to get container status \"fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44\": rpc error: code = NotFound desc = could not find container \"fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44\": container with ID starting with fe4bd447e5b5830799222164a24fa8f88eccd490ced5787bcbd45ca57bdd4d44 not found: ID does not exist" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.895945 4799 scope.go:117] "RemoveContainer" containerID="d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7" Sep 30 14:44:04 crc kubenswrapper[4799]: E0930 14:44:04.896561 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7\": container with ID starting with d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7 not found: ID does not exist" containerID="d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.896599 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7"} err="failed to get container status \"d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7\": rpc error: code = NotFound desc = could not find container \"d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7\": container with ID starting with d9113196b2bdbc87215d5d7093067bb990b3c034b23670f4f46c9ee2e0ac2ea7 not found: ID does not exist" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.928566 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce901e13-dfa1-4b79-a584-4903008803d0-logs\") pod \"ce901e13-dfa1-4b79-a584-4903008803d0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.928800 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-nova-metadata-tls-certs\") pod \"ce901e13-dfa1-4b79-a584-4903008803d0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.928867 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqdrt\" (UniqueName: \"kubernetes.io/projected/ce901e13-dfa1-4b79-a584-4903008803d0-kube-api-access-zqdrt\") pod \"ce901e13-dfa1-4b79-a584-4903008803d0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.928931 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-config-data\") pod \"ce901e13-dfa1-4b79-a584-4903008803d0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.928998 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-combined-ca-bundle\") pod \"ce901e13-dfa1-4b79-a584-4903008803d0\" (UID: \"ce901e13-dfa1-4b79-a584-4903008803d0\") " Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.929416 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce901e13-dfa1-4b79-a584-4903008803d0-logs" (OuterVolumeSpecName: "logs") pod "ce901e13-dfa1-4b79-a584-4903008803d0" (UID: "ce901e13-dfa1-4b79-a584-4903008803d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.943874 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce901e13-dfa1-4b79-a584-4903008803d0-kube-api-access-zqdrt" (OuterVolumeSpecName: "kube-api-access-zqdrt") pod "ce901e13-dfa1-4b79-a584-4903008803d0" (UID: "ce901e13-dfa1-4b79-a584-4903008803d0"). InnerVolumeSpecName "kube-api-access-zqdrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:44:04 crc kubenswrapper[4799]: I0930 14:44:04.970725 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce901e13-dfa1-4b79-a584-4903008803d0" (UID: "ce901e13-dfa1-4b79-a584-4903008803d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.007803 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-config-data" (OuterVolumeSpecName: "config-data") pod "ce901e13-dfa1-4b79-a584-4903008803d0" (UID: "ce901e13-dfa1-4b79-a584-4903008803d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.014227 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ce901e13-dfa1-4b79-a584-4903008803d0" (UID: "ce901e13-dfa1-4b79-a584-4903008803d0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.031486 4799 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.031543 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqdrt\" (UniqueName: \"kubernetes.io/projected/ce901e13-dfa1-4b79-a584-4903008803d0-kube-api-access-zqdrt\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.031557 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.031573 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce901e13-dfa1-4b79-a584-4903008803d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.031584 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce901e13-dfa1-4b79-a584-4903008803d0-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.322570 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.360843 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.398098 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.398707 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="extract-utilities" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.398735 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="extract-utilities" Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.398787 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="extract-content" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.398797 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="extract-content" Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.398816 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-metadata" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.398824 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-metadata" Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.398861 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-log" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.398869 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-log" Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.398887 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="registry-server" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.398895 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="registry-server" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.399170 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-log" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.399187 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" containerName="nova-metadata-metadata" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.399205 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f6380d4-aaf5-41ac-b134-9f46269f2183" containerName="registry-server" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.400578 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.410999 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.411024 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.411205 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.464224 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04 is running failed: container process not found" containerID="21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.466241 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04 is running failed: container process not found" containerID="21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.466639 4799 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04 is running failed: container process not found" containerID="21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.466800 4799 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d7a4ded6-46db-4145-a518-f9b3dee3843e" containerName="nova-scheduler-scheduler" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.547273 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.547376 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j2b8\" (UniqueName: \"kubernetes.io/projected/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-kube-api-access-8j2b8\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.547413 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.547453 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-logs\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.547515 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-config-data\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.650205 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.650395 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j2b8\" (UniqueName: \"kubernetes.io/projected/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-kube-api-access-8j2b8\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.650440 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.650494 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-logs\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.650609 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-config-data\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.652487 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-logs\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.659394 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.659460 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.661532 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-config-data\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.665159 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.682893 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j2b8\" (UniqueName: \"kubernetes.io/projected/fad4220f-11f3-45e3-b5f5-70d210ad5f3f-kube-api-access-8j2b8\") pod \"nova-metadata-0\" (UID: \"fad4220f-11f3-45e3-b5f5-70d210ad5f3f\") " pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.723333 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.855458 4799 generic.go:334] "Generic (PLEG): container finished" podID="d7a4ded6-46db-4145-a518-f9b3dee3843e" containerID="21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04" exitCode=0 Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.857340 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.857925 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7a4ded6-46db-4145-a518-f9b3dee3843e","Type":"ContainerDied","Data":"21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04"} Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.858008 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7a4ded6-46db-4145-a518-f9b3dee3843e","Type":"ContainerDied","Data":"9f55b8b6a4452906ff32bcc330a55b8e2842445e86ca741ea7fe3bebed621b70"} Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.858041 4799 scope.go:117] "RemoveContainer" containerID="21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.858265 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-combined-ca-bundle\") pod \"d7a4ded6-46db-4145-a518-f9b3dee3843e\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.858439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n42fl\" (UniqueName: \"kubernetes.io/projected/d7a4ded6-46db-4145-a518-f9b3dee3843e-kube-api-access-n42fl\") pod \"d7a4ded6-46db-4145-a518-f9b3dee3843e\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.858597 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-config-data\") pod \"d7a4ded6-46db-4145-a518-f9b3dee3843e\" (UID: \"d7a4ded6-46db-4145-a518-f9b3dee3843e\") " Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.867618 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a4ded6-46db-4145-a518-f9b3dee3843e-kube-api-access-n42fl" (OuterVolumeSpecName: "kube-api-access-n42fl") pod "d7a4ded6-46db-4145-a518-f9b3dee3843e" (UID: "d7a4ded6-46db-4145-a518-f9b3dee3843e"). InnerVolumeSpecName "kube-api-access-n42fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.901751 4799 scope.go:117] "RemoveContainer" containerID="21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04" Sep 30 14:44:05 crc kubenswrapper[4799]: E0930 14:44:05.902351 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04\": container with ID starting with 21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04 not found: ID does not exist" containerID="21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.902392 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04"} err="failed to get container status \"21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04\": rpc error: code = NotFound desc = could not find container \"21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04\": container with ID starting with 21558b60c8b0bb20a77f7b10870cdb5e45b227430e358e91edec26369bebdd04 not found: ID does not exist" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.913681 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7a4ded6-46db-4145-a518-f9b3dee3843e" (UID: "d7a4ded6-46db-4145-a518-f9b3dee3843e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.939969 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-config-data" (OuterVolumeSpecName: "config-data") pod "d7a4ded6-46db-4145-a518-f9b3dee3843e" (UID: "d7a4ded6-46db-4145-a518-f9b3dee3843e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.976179 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n42fl\" (UniqueName: \"kubernetes.io/projected/d7a4ded6-46db-4145-a518-f9b3dee3843e-kube-api-access-n42fl\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.976218 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:05 crc kubenswrapper[4799]: I0930 14:44:05.976232 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a4ded6-46db-4145-a518-f9b3dee3843e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.248518 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.263962 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.278683 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:44:06 crc kubenswrapper[4799]: E0930 14:44:06.279207 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a4ded6-46db-4145-a518-f9b3dee3843e" containerName="nova-scheduler-scheduler" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.279235 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a4ded6-46db-4145-a518-f9b3dee3843e" containerName="nova-scheduler-scheduler" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.279534 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a4ded6-46db-4145-a518-f9b3dee3843e" containerName="nova-scheduler-scheduler" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.283960 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.291576 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.312685 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.328969 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.384796 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/886bc5b6-844e-4a31-8fe2-d62c56108e3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.384972 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7b5t\" (UniqueName: \"kubernetes.io/projected/886bc5b6-844e-4a31-8fe2-d62c56108e3a-kube-api-access-v7b5t\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.385011 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/886bc5b6-844e-4a31-8fe2-d62c56108e3a-config-data\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.497548 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7b5t\" (UniqueName: \"kubernetes.io/projected/886bc5b6-844e-4a31-8fe2-d62c56108e3a-kube-api-access-v7b5t\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.497888 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/886bc5b6-844e-4a31-8fe2-d62c56108e3a-config-data\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.498011 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/886bc5b6-844e-4a31-8fe2-d62c56108e3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.504560 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/886bc5b6-844e-4a31-8fe2-d62c56108e3a-config-data\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.507690 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/886bc5b6-844e-4a31-8fe2-d62c56108e3a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.534905 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce901e13-dfa1-4b79-a584-4903008803d0" path="/var/lib/kubelet/pods/ce901e13-dfa1-4b79-a584-4903008803d0/volumes" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.537585 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7b5t\" (UniqueName: \"kubernetes.io/projected/886bc5b6-844e-4a31-8fe2-d62c56108e3a-kube-api-access-v7b5t\") pod \"nova-scheduler-0\" (UID: \"886bc5b6-844e-4a31-8fe2-d62c56108e3a\") " pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.541302 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7a4ded6-46db-4145-a518-f9b3dee3843e" path="/var/lib/kubelet/pods/d7a4ded6-46db-4145-a518-f9b3dee3843e/volumes" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.724950 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.738044 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.897901 4799 generic.go:334] "Generic (PLEG): container finished" podID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerID="bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736" exitCode=0 Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.897991 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerDied","Data":"bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736"} Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.898039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"77b5c423-17cd-416f-91cf-56aa87947c5c","Type":"ContainerDied","Data":"331947af2a2d948861686f78dac43e8c8292bf5feb4677d3c4107e98c970c881"} Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.898085 4799 scope.go:117] "RemoveContainer" containerID="2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.898390 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.924009 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-sg-core-conf-yaml\") pod \"77b5c423-17cd-416f-91cf-56aa87947c5c\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.924084 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6spnh\" (UniqueName: \"kubernetes.io/projected/77b5c423-17cd-416f-91cf-56aa87947c5c-kube-api-access-6spnh\") pod \"77b5c423-17cd-416f-91cf-56aa87947c5c\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.924142 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-log-httpd\") pod \"77b5c423-17cd-416f-91cf-56aa87947c5c\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.924241 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-ceilometer-tls-certs\") pod \"77b5c423-17cd-416f-91cf-56aa87947c5c\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.924362 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-scripts\") pod \"77b5c423-17cd-416f-91cf-56aa87947c5c\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.924442 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-combined-ca-bundle\") pod \"77b5c423-17cd-416f-91cf-56aa87947c5c\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.924484 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-config-data\") pod \"77b5c423-17cd-416f-91cf-56aa87947c5c\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.924561 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-run-httpd\") pod \"77b5c423-17cd-416f-91cf-56aa87947c5c\" (UID: \"77b5c423-17cd-416f-91cf-56aa87947c5c\") " Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.929418 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77b5c423-17cd-416f-91cf-56aa87947c5c-kube-api-access-6spnh" (OuterVolumeSpecName: "kube-api-access-6spnh") pod "77b5c423-17cd-416f-91cf-56aa87947c5c" (UID: "77b5c423-17cd-416f-91cf-56aa87947c5c"). InnerVolumeSpecName "kube-api-access-6spnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.930809 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "77b5c423-17cd-416f-91cf-56aa87947c5c" (UID: "77b5c423-17cd-416f-91cf-56aa87947c5c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.934075 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "77b5c423-17cd-416f-91cf-56aa87947c5c" (UID: "77b5c423-17cd-416f-91cf-56aa87947c5c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.935587 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fad4220f-11f3-45e3-b5f5-70d210ad5f3f","Type":"ContainerStarted","Data":"54317a38d4693617a65e4bb6ea59f67c81c680206ddd0a69056bdd89f954f7a8"} Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.935665 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fad4220f-11f3-45e3-b5f5-70d210ad5f3f","Type":"ContainerStarted","Data":"c068f69a8c96959e30f0d43d5ce3a5ab91147aec093c0a8a88db11bd22e79130"} Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.965693 4799 scope.go:117] "RemoveContainer" containerID="e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c" Sep 30 14:44:06 crc kubenswrapper[4799]: I0930 14:44:06.986082 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-scripts" (OuterVolumeSpecName: "scripts") pod "77b5c423-17cd-416f-91cf-56aa87947c5c" (UID: "77b5c423-17cd-416f-91cf-56aa87947c5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.012047 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "77b5c423-17cd-416f-91cf-56aa87947c5c" (UID: "77b5c423-17cd-416f-91cf-56aa87947c5c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.028153 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6spnh\" (UniqueName: \"kubernetes.io/projected/77b5c423-17cd-416f-91cf-56aa87947c5c-kube-api-access-6spnh\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.028196 4799 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-log-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.028208 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.028222 4799 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/77b5c423-17cd-416f-91cf-56aa87947c5c-run-httpd\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.028230 4799 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.050605 4799 scope.go:117] "RemoveContainer" containerID="84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.083750 4799 scope.go:117] "RemoveContainer" containerID="bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.114548 4799 scope.go:117] "RemoveContainer" containerID="2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402" Sep 30 14:44:07 crc kubenswrapper[4799]: E0930 14:44:07.115582 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402\": container with ID starting with 2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402 not found: ID does not exist" containerID="2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.115686 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402"} err="failed to get container status \"2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402\": rpc error: code = NotFound desc = could not find container \"2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402\": container with ID starting with 2f39f4d165cb20a2a5ac623f55b8a307471d11ed1ebe3ec7af45d98025b02402 not found: ID does not exist" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.115764 4799 scope.go:117] "RemoveContainer" containerID="e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c" Sep 30 14:44:07 crc kubenswrapper[4799]: E0930 14:44:07.116233 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c\": container with ID starting with e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c not found: ID does not exist" containerID="e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.116268 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c"} err="failed to get container status \"e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c\": rpc error: code = NotFound desc = could not find container \"e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c\": container with ID starting with e5c87a359539237100ab642d1759a28f07cc41ed540de02ff2f7388984249c8c not found: ID does not exist" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.116302 4799 scope.go:117] "RemoveContainer" containerID="84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73" Sep 30 14:44:07 crc kubenswrapper[4799]: E0930 14:44:07.116806 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73\": container with ID starting with 84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73 not found: ID does not exist" containerID="84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.116838 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73"} err="failed to get container status \"84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73\": rpc error: code = NotFound desc = could not find container \"84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73\": container with ID starting with 84be7512cc2b93a80c7d8f336381de93fb139dfa3c71ab79df85575c9ed93d73 not found: ID does not exist" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.116867 4799 scope.go:117] "RemoveContainer" containerID="bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736" Sep 30 14:44:07 crc kubenswrapper[4799]: E0930 14:44:07.117190 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736\": container with ID starting with bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736 not found: ID does not exist" containerID="bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.117269 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736"} err="failed to get container status \"bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736\": rpc error: code = NotFound desc = could not find container \"bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736\": container with ID starting with bb5e69f17a277b7d635c699270d28063a184170786475beeeec667751b512736 not found: ID does not exist" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.117895 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "77b5c423-17cd-416f-91cf-56aa87947c5c" (UID: "77b5c423-17cd-416f-91cf-56aa87947c5c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.133558 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.188427 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77b5c423-17cd-416f-91cf-56aa87947c5c" (UID: "77b5c423-17cd-416f-91cf-56aa87947c5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.226041 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-config-data" (OuterVolumeSpecName: "config-data") pod "77b5c423-17cd-416f-91cf-56aa87947c5c" (UID: "77b5c423-17cd-416f-91cf-56aa87947c5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.235881 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.236595 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77b5c423-17cd-416f-91cf-56aa87947c5c-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.341089 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.529504 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.529628 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.530843 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"e36dbd20d5d3cdf7ece15c8b2611df5884ae476a6712a79abbb5e85a1de0277d"} pod="openstack/horizon-7f99689854-glm7t" containerMessage="Container horizon failed startup probe, will be restarted" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.530899 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" containerID="cri-o://e36dbd20d5d3cdf7ece15c8b2611df5884ae476a6712a79abbb5e85a1de0277d" gracePeriod=30 Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.558990 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.591795 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.628668 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:44:07 crc kubenswrapper[4799]: E0930 14:44:07.629276 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="ceilometer-central-agent" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.629294 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="ceilometer-central-agent" Sep 30 14:44:07 crc kubenswrapper[4799]: E0930 14:44:07.629318 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="proxy-httpd" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.629326 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="proxy-httpd" Sep 30 14:44:07 crc kubenswrapper[4799]: E0930 14:44:07.629347 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="ceilometer-notification-agent" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.629353 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="ceilometer-notification-agent" Sep 30 14:44:07 crc kubenswrapper[4799]: E0930 14:44:07.629371 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="sg-core" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.629376 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="sg-core" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.629578 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="ceilometer-notification-agent" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.629596 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="proxy-httpd" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.629603 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="sg-core" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.629614 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" containerName="ceilometer-central-agent" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.639046 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.639204 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.658869 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.659119 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.659143 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.766272 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.766341 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.766576 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-config-data\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.766710 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-log-httpd\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.766793 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hclf\" (UniqueName: \"kubernetes.io/projected/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-kube-api-access-6hclf\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.767133 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.767190 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-scripts\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.767268 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-run-httpd\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.869740 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-scripts\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.869813 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-run-httpd\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.869887 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.869928 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.869980 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-config-data\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.870018 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-log-httpd\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.870041 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hclf\" (UniqueName: \"kubernetes.io/projected/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-kube-api-access-6hclf\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.870132 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.876392 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.876891 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.879989 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-run-httpd\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.880596 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-scripts\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.881433 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-log-httpd\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.882399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-config-data\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.886902 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.902804 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hclf\" (UniqueName: \"kubernetes.io/projected/6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a-kube-api-access-6hclf\") pod \"ceilometer-0\" (UID: \"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a\") " pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.960629 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fad4220f-11f3-45e3-b5f5-70d210ad5f3f","Type":"ContainerStarted","Data":"dbd3007e7fd1c55b7bcee84a03abc91d289f32aff1abe871503685d8fdff5800"} Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.964905 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.968232 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"886bc5b6-844e-4a31-8fe2-d62c56108e3a","Type":"ContainerStarted","Data":"37330772941b88de39300c213eadefaa969b32f46de41c6c477079ee4c0e2121"} Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.968277 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"886bc5b6-844e-4a31-8fe2-d62c56108e3a","Type":"ContainerStarted","Data":"da4771890a3ae4bd3c5b683c07a285f0302a2352ea9a8d82443515e5e5004023"} Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.971960 4799 generic.go:334] "Generic (PLEG): container finished" podID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerID="7630e75f64223c31df72c9c0ad92ebccc462f64abfa85db6dbbf91edb9c7433c" exitCode=137 Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.972023 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerDied","Data":"7630e75f64223c31df72c9c0ad92ebccc462f64abfa85db6dbbf91edb9c7433c"} Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.972063 4799 scope.go:117] "RemoveContainer" containerID="145eff7bbb6dd0a52fa1de14ab2ef3666affe0c5dae478a2644a1c0dffeb1113" Sep 30 14:44:07 crc kubenswrapper[4799]: I0930 14:44:07.987702 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.987679082 podStartE2EDuration="2.987679082s" podCreationTimestamp="2025-09-30 14:44:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:44:07.986343463 +0000 UTC m=+1470.069943890" watchObservedRunningTime="2025-09-30 14:44:07.987679082 +0000 UTC m=+1470.071279509" Sep 30 14:44:08 crc kubenswrapper[4799]: I0930 14:44:08.019616 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.019592625 podStartE2EDuration="2.019592625s" podCreationTimestamp="2025-09-30 14:44:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:44:08.009136799 +0000 UTC m=+1470.092737256" watchObservedRunningTime="2025-09-30 14:44:08.019592625 +0000 UTC m=+1470.103193052" Sep 30 14:44:08 crc kubenswrapper[4799]: I0930 14:44:08.516267 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77b5c423-17cd-416f-91cf-56aa87947c5c" path="/var/lib/kubelet/pods/77b5c423-17cd-416f-91cf-56aa87947c5c/volumes" Sep 30 14:44:08 crc kubenswrapper[4799]: W0930 14:44:08.572480 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dcb5a2b_4d03_4b9f_a92a_4ee965f1502a.slice/crio-b3fe1d190be6849dcb4eb4e1cc64844819b353bcc4baae24d8ddc8cc0ca02798 WatchSource:0}: Error finding container b3fe1d190be6849dcb4eb4e1cc64844819b353bcc4baae24d8ddc8cc0ca02798: Status 404 returned error can't find the container with id b3fe1d190be6849dcb4eb4e1cc64844819b353bcc4baae24d8ddc8cc0ca02798 Sep 30 14:44:08 crc kubenswrapper[4799]: I0930 14:44:08.574503 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Sep 30 14:44:08 crc kubenswrapper[4799]: I0930 14:44:08.984721 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a","Type":"ContainerStarted","Data":"b3fe1d190be6849dcb4eb4e1cc64844819b353bcc4baae24d8ddc8cc0ca02798"} Sep 30 14:44:08 crc kubenswrapper[4799]: I0930 14:44:08.990317 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f7c888686-pwv2q" event={"ID":"bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6","Type":"ContainerStarted","Data":"ebe5a5851d4ba3f1592e03c700d4469752b307d18d355c3041cab4978633148b"} Sep 30 14:44:10 crc kubenswrapper[4799]: I0930 14:44:10.041287 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a","Type":"ContainerStarted","Data":"d62371acf1e139b6b29460a37e7a807fce2ee9c0e59b0578dfea090ca1e74ca7"} Sep 30 14:44:10 crc kubenswrapper[4799]: I0930 14:44:10.723723 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:44:10 crc kubenswrapper[4799]: I0930 14:44:10.724247 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Sep 30 14:44:11 crc kubenswrapper[4799]: I0930 14:44:11.061417 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a","Type":"ContainerStarted","Data":"83fea072f8ce36ba0dcb414949d9f7f61658d628938c8b54188d1893a5a189d6"} Sep 30 14:44:11 crc kubenswrapper[4799]: I0930 14:44:11.061473 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a","Type":"ContainerStarted","Data":"d7b50f982a3d808e9b6c65177d80bb673b9d2e753019b62e7e2ab77ae66f7d23"} Sep 30 14:44:11 crc kubenswrapper[4799]: I0930 14:44:11.739359 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Sep 30 14:44:13 crc kubenswrapper[4799]: I0930 14:44:13.085858 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a","Type":"ContainerStarted","Data":"9dcf1f18b8d954f70148f0794d279e1789fc67a177d0f93421da36e0c088e460"} Sep 30 14:44:13 crc kubenswrapper[4799]: I0930 14:44:13.086396 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Sep 30 14:44:13 crc kubenswrapper[4799]: I0930 14:44:13.129752 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.440061709 podStartE2EDuration="6.12962404s" podCreationTimestamp="2025-09-30 14:44:07 +0000 UTC" firstStartedPulling="2025-09-30 14:44:08.577933347 +0000 UTC m=+1470.661533774" lastFinishedPulling="2025-09-30 14:44:12.267495678 +0000 UTC m=+1474.351096105" observedRunningTime="2025-09-30 14:44:13.118730252 +0000 UTC m=+1475.202330689" watchObservedRunningTime="2025-09-30 14:44:13.12962404 +0000 UTC m=+1475.213224467" Sep 30 14:44:13 crc kubenswrapper[4799]: I0930 14:44:13.342840 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:44:13 crc kubenswrapper[4799]: I0930 14:44:13.342932 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Sep 30 14:44:14 crc kubenswrapper[4799]: I0930 14:44:14.367481 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="28bc13f2-5895-42cf-8ab5-60251eb4d09f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:44:14 crc kubenswrapper[4799]: I0930 14:44:14.370209 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="28bc13f2-5895-42cf-8ab5-60251eb4d09f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:44:15 crc kubenswrapper[4799]: I0930 14:44:15.724450 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 14:44:15 crc kubenswrapper[4799]: I0930 14:44:15.724891 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Sep 30 14:44:16 crc kubenswrapper[4799]: I0930 14:44:16.738271 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fad4220f-11f3-45e3-b5f5-70d210ad5f3f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 30 14:44:16 crc kubenswrapper[4799]: I0930 14:44:16.738731 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fad4220f-11f3-45e3-b5f5-70d210ad5f3f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 30 14:44:16 crc kubenswrapper[4799]: I0930 14:44:16.738900 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Sep 30 14:44:16 crc kubenswrapper[4799]: I0930 14:44:16.779761 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Sep 30 14:44:17 crc kubenswrapper[4799]: I0930 14:44:17.177196 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Sep 30 14:44:17 crc kubenswrapper[4799]: I0930 14:44:17.681344 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:44:17 crc kubenswrapper[4799]: I0930 14:44:17.681630 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:44:23 crc kubenswrapper[4799]: I0930 14:44:23.352106 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:44:23 crc kubenswrapper[4799]: I0930 14:44:23.352697 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Sep 30 14:44:23 crc kubenswrapper[4799]: I0930 14:44:23.353140 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:44:23 crc kubenswrapper[4799]: I0930 14:44:23.353603 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Sep 30 14:44:23 crc kubenswrapper[4799]: I0930 14:44:23.374568 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:44:23 crc kubenswrapper[4799]: I0930 14:44:23.379892 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Sep 30 14:44:25 crc kubenswrapper[4799]: I0930 14:44:25.730579 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 14:44:25 crc kubenswrapper[4799]: I0930 14:44:25.731342 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Sep 30 14:44:25 crc kubenswrapper[4799]: I0930 14:44:25.744169 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 14:44:25 crc kubenswrapper[4799]: I0930 14:44:25.744224 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Sep 30 14:44:27 crc kubenswrapper[4799]: I0930 14:44:27.682566 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f7c888686-pwv2q" podUID="bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.136274 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l4lgb"] Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.141020 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.151721 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4lgb"] Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.288954 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6qgp\" (UniqueName: \"kubernetes.io/projected/62e47d05-f061-496c-ac37-47ff40532ee2-kube-api-access-b6qgp\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.289099 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-utilities\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.289306 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-catalog-content\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.391619 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6qgp\" (UniqueName: \"kubernetes.io/projected/62e47d05-f061-496c-ac37-47ff40532ee2-kube-api-access-b6qgp\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.391839 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-utilities\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.391877 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-catalog-content\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.393213 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-catalog-content\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.393927 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-utilities\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.417197 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6qgp\" (UniqueName: \"kubernetes.io/projected/62e47d05-f061-496c-ac37-47ff40532ee2-kube-api-access-b6qgp\") pod \"redhat-marketplace-l4lgb\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.485694 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.649324 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:44:29 crc kubenswrapper[4799]: I0930 14:44:29.649649 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:44:30 crc kubenswrapper[4799]: I0930 14:44:30.025245 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4lgb"] Sep 30 14:44:30 crc kubenswrapper[4799]: I0930 14:44:30.293241 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4lgb" event={"ID":"62e47d05-f061-496c-ac37-47ff40532ee2","Type":"ContainerStarted","Data":"303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063"} Sep 30 14:44:30 crc kubenswrapper[4799]: I0930 14:44:30.293309 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4lgb" event={"ID":"62e47d05-f061-496c-ac37-47ff40532ee2","Type":"ContainerStarted","Data":"458f8e24b93aaab14ca65f152b3d0ebe0954eed744b08171b326344bacf19fad"} Sep 30 14:44:31 crc kubenswrapper[4799]: I0930 14:44:31.314051 4799 generic.go:334] "Generic (PLEG): container finished" podID="62e47d05-f061-496c-ac37-47ff40532ee2" containerID="303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063" exitCode=0 Sep 30 14:44:31 crc kubenswrapper[4799]: I0930 14:44:31.314446 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4lgb" event={"ID":"62e47d05-f061-496c-ac37-47ff40532ee2","Type":"ContainerDied","Data":"303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063"} Sep 30 14:44:32 crc kubenswrapper[4799]: I0930 14:44:32.332313 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4lgb" event={"ID":"62e47d05-f061-496c-ac37-47ff40532ee2","Type":"ContainerStarted","Data":"b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac"} Sep 30 14:44:33 crc kubenswrapper[4799]: I0930 14:44:33.349844 4799 generic.go:334] "Generic (PLEG): container finished" podID="62e47d05-f061-496c-ac37-47ff40532ee2" containerID="b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac" exitCode=0 Sep 30 14:44:33 crc kubenswrapper[4799]: I0930 14:44:33.349920 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4lgb" event={"ID":"62e47d05-f061-496c-ac37-47ff40532ee2","Type":"ContainerDied","Data":"b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac"} Sep 30 14:44:34 crc kubenswrapper[4799]: I0930 14:44:34.373388 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4lgb" event={"ID":"62e47d05-f061-496c-ac37-47ff40532ee2","Type":"ContainerStarted","Data":"eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b"} Sep 30 14:44:34 crc kubenswrapper[4799]: I0930 14:44:34.398199 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l4lgb" podStartSLOduration=2.684025498 podStartE2EDuration="5.398159403s" podCreationTimestamp="2025-09-30 14:44:29 +0000 UTC" firstStartedPulling="2025-09-30 14:44:31.321925083 +0000 UTC m=+1493.405525510" lastFinishedPulling="2025-09-30 14:44:34.036058998 +0000 UTC m=+1496.119659415" observedRunningTime="2025-09-30 14:44:34.39636714 +0000 UTC m=+1496.479967587" watchObservedRunningTime="2025-09-30 14:44:34.398159403 +0000 UTC m=+1496.481759830" Sep 30 14:44:37 crc kubenswrapper[4799]: I0930 14:44:37.984734 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Sep 30 14:44:38 crc kubenswrapper[4799]: I0930 14:44:38.421696 4799 generic.go:334] "Generic (PLEG): container finished" podID="d15890bb-d088-4126-8382-b5d8e423b608" containerID="e36dbd20d5d3cdf7ece15c8b2611df5884ae476a6712a79abbb5e85a1de0277d" exitCode=137 Sep 30 14:44:38 crc kubenswrapper[4799]: I0930 14:44:38.421783 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerDied","Data":"e36dbd20d5d3cdf7ece15c8b2611df5884ae476a6712a79abbb5e85a1de0277d"} Sep 30 14:44:38 crc kubenswrapper[4799]: I0930 14:44:38.422082 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerStarted","Data":"15813eeea4cf4efe8492111998588d8903edae8edcbbf0a1ab40735200a85876"} Sep 30 14:44:38 crc kubenswrapper[4799]: I0930 14:44:38.422116 4799 scope.go:117] "RemoveContainer" containerID="6a24eead33c55554d4fa163faed09d5db7e530d41a8e1d35cb03d928a56c1750" Sep 30 14:44:39 crc kubenswrapper[4799]: I0930 14:44:39.486178 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:39 crc kubenswrapper[4799]: I0930 14:44:39.486549 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:39 crc kubenswrapper[4799]: I0930 14:44:39.557972 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:40 crc kubenswrapper[4799]: I0930 14:44:40.529462 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:40 crc kubenswrapper[4799]: I0930 14:44:40.591626 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4lgb"] Sep 30 14:44:41 crc kubenswrapper[4799]: I0930 14:44:41.765338 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:44:42 crc kubenswrapper[4799]: I0930 14:44:42.490424 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l4lgb" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" containerName="registry-server" containerID="cri-o://eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b" gracePeriod=2 Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.010585 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.095512 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-utilities\") pod \"62e47d05-f061-496c-ac37-47ff40532ee2\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.095982 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6qgp\" (UniqueName: \"kubernetes.io/projected/62e47d05-f061-496c-ac37-47ff40532ee2-kube-api-access-b6qgp\") pod \"62e47d05-f061-496c-ac37-47ff40532ee2\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.096181 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-catalog-content\") pod \"62e47d05-f061-496c-ac37-47ff40532ee2\" (UID: \"62e47d05-f061-496c-ac37-47ff40532ee2\") " Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.096431 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-utilities" (OuterVolumeSpecName: "utilities") pod "62e47d05-f061-496c-ac37-47ff40532ee2" (UID: "62e47d05-f061-496c-ac37-47ff40532ee2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.097001 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.103393 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e47d05-f061-496c-ac37-47ff40532ee2-kube-api-access-b6qgp" (OuterVolumeSpecName: "kube-api-access-b6qgp") pod "62e47d05-f061-496c-ac37-47ff40532ee2" (UID: "62e47d05-f061-496c-ac37-47ff40532ee2"). InnerVolumeSpecName "kube-api-access-b6qgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.113868 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62e47d05-f061-496c-ac37-47ff40532ee2" (UID: "62e47d05-f061-496c-ac37-47ff40532ee2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.198958 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6qgp\" (UniqueName: \"kubernetes.io/projected/62e47d05-f061-496c-ac37-47ff40532ee2-kube-api-access-b6qgp\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.199014 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e47d05-f061-496c-ac37-47ff40532ee2-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.504920 4799 generic.go:334] "Generic (PLEG): container finished" podID="62e47d05-f061-496c-ac37-47ff40532ee2" containerID="eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b" exitCode=0 Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.504971 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4lgb" event={"ID":"62e47d05-f061-496c-ac37-47ff40532ee2","Type":"ContainerDied","Data":"eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b"} Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.505007 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4lgb" event={"ID":"62e47d05-f061-496c-ac37-47ff40532ee2","Type":"ContainerDied","Data":"458f8e24b93aaab14ca65f152b3d0ebe0954eed744b08171b326344bacf19fad"} Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.505034 4799 scope.go:117] "RemoveContainer" containerID="eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.505480 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4lgb" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.539257 4799 scope.go:117] "RemoveContainer" containerID="b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.549699 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4lgb"] Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.561448 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4lgb"] Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.577943 4799 scope.go:117] "RemoveContainer" containerID="303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.621672 4799 scope.go:117] "RemoveContainer" containerID="eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b" Sep 30 14:44:43 crc kubenswrapper[4799]: E0930 14:44:43.622186 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b\": container with ID starting with eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b not found: ID does not exist" containerID="eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.622353 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b"} err="failed to get container status \"eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b\": rpc error: code = NotFound desc = could not find container \"eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b\": container with ID starting with eb4ac213906606880e962ce3f8ed532c0781194049cc155450573cf9c81f3a8b not found: ID does not exist" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.622445 4799 scope.go:117] "RemoveContainer" containerID="b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac" Sep 30 14:44:43 crc kubenswrapper[4799]: E0930 14:44:43.623160 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac\": container with ID starting with b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac not found: ID does not exist" containerID="b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.623189 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac"} err="failed to get container status \"b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac\": rpc error: code = NotFound desc = could not find container \"b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac\": container with ID starting with b2a0ba26d0a0080d0611b129292037c9eeb9f539b6d5fe4bc2fba0f2c9e5d2ac not found: ID does not exist" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.623207 4799 scope.go:117] "RemoveContainer" containerID="303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063" Sep 30 14:44:43 crc kubenswrapper[4799]: E0930 14:44:43.623489 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063\": container with ID starting with 303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063 not found: ID does not exist" containerID="303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.623513 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063"} err="failed to get container status \"303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063\": rpc error: code = NotFound desc = could not find container \"303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063\": container with ID starting with 303769bc29f5613fc0a8902af53e74e1ef9e34c14d18dc48417dadd7c57c0063 not found: ID does not exist" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.922852 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7f7c888686-pwv2q" Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.998559 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f99689854-glm7t"] Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.999265 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon-log" containerID="cri-o://8923b064d4910eddccd36258c0091f291d87f5d2b5ba8a2a2855d3bafc5f9665" gracePeriod=30 Sep 30 14:44:43 crc kubenswrapper[4799]: I0930 14:44:43.999479 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f99689854-glm7t" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" containerID="cri-o://15813eeea4cf4efe8492111998588d8903edae8edcbbf0a1ab40735200a85876" gracePeriod=30 Sep 30 14:44:44 crc kubenswrapper[4799]: I0930 14:44:44.517817 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" path="/var/lib/kubelet/pods/62e47d05-f061-496c-ac37-47ff40532ee2/volumes" Sep 30 14:44:47 crc kubenswrapper[4799]: I0930 14:44:47.528771 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:44:53 crc kubenswrapper[4799]: I0930 14:44:53.510978 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:44:54 crc kubenswrapper[4799]: I0930 14:44:54.820099 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:44:59 crc kubenswrapper[4799]: I0930 14:44:59.649151 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:44:59 crc kubenswrapper[4799]: I0930 14:44:59.649793 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.197498 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49"] Sep 30 14:45:00 crc kubenswrapper[4799]: E0930 14:45:00.198207 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" containerName="extract-utilities" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.198234 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" containerName="extract-utilities" Sep 30 14:45:00 crc kubenswrapper[4799]: E0930 14:45:00.198265 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" containerName="extract-content" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.198274 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" containerName="extract-content" Sep 30 14:45:00 crc kubenswrapper[4799]: E0930 14:45:00.198312 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" containerName="registry-server" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.198319 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" containerName="registry-server" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.198597 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e47d05-f061-496c-ac37-47ff40532ee2" containerName="registry-server" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.202992 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.207330 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.211674 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.249835 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49"] Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.311787 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e013782-a27d-4cf6-bc7c-29a13bbd4137-config-volume\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.312525 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbzmh\" (UniqueName: \"kubernetes.io/projected/1e013782-a27d-4cf6-bc7c-29a13bbd4137-kube-api-access-pbzmh\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.312729 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e013782-a27d-4cf6-bc7c-29a13bbd4137-secret-volume\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.415342 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e013782-a27d-4cf6-bc7c-29a13bbd4137-secret-volume\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.415992 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e013782-a27d-4cf6-bc7c-29a13bbd4137-config-volume\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.416134 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbzmh\" (UniqueName: \"kubernetes.io/projected/1e013782-a27d-4cf6-bc7c-29a13bbd4137-kube-api-access-pbzmh\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.417511 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e013782-a27d-4cf6-bc7c-29a13bbd4137-config-volume\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.429715 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e013782-a27d-4cf6-bc7c-29a13bbd4137-secret-volume\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.441422 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbzmh\" (UniqueName: \"kubernetes.io/projected/1e013782-a27d-4cf6-bc7c-29a13bbd4137-kube-api-access-pbzmh\") pod \"collect-profiles-29320725-p9j49\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.547608 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.812272 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerName="rabbitmq" containerID="cri-o://d0a099467cee71f3c37c55698919e290d84a71960fad48a65565f735a6ca095f" gracePeriod=604793 Sep 30 14:45:00 crc kubenswrapper[4799]: I0930 14:45:00.996452 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerName="rabbitmq" containerID="cri-o://d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7" gracePeriod=604794 Sep 30 14:45:01 crc kubenswrapper[4799]: I0930 14:45:01.597887 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49"] Sep 30 14:45:01 crc kubenswrapper[4799]: I0930 14:45:01.776222 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" event={"ID":"1e013782-a27d-4cf6-bc7c-29a13bbd4137","Type":"ContainerStarted","Data":"cd0e0b8af9898ed25e09ab5abd2e2899153dc5dc0b232d01e218e519a4a0d7f6"} Sep 30 14:45:02 crc kubenswrapper[4799]: I0930 14:45:02.788147 4799 generic.go:334] "Generic (PLEG): container finished" podID="1e013782-a27d-4cf6-bc7c-29a13bbd4137" containerID="cb13ec04eb93bc6be8b6a7ea9c48dddfcbd4ca4e489636300c5b03a962ff2961" exitCode=0 Sep 30 14:45:02 crc kubenswrapper[4799]: I0930 14:45:02.788264 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" event={"ID":"1e013782-a27d-4cf6-bc7c-29a13bbd4137","Type":"ContainerDied","Data":"cb13ec04eb93bc6be8b6a7ea9c48dddfcbd4ca4e489636300c5b03a962ff2961"} Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.403019 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.510598 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e013782-a27d-4cf6-bc7c-29a13bbd4137-secret-volume\") pod \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.511424 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e013782-a27d-4cf6-bc7c-29a13bbd4137-config-volume\") pod \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.511598 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbzmh\" (UniqueName: \"kubernetes.io/projected/1e013782-a27d-4cf6-bc7c-29a13bbd4137-kube-api-access-pbzmh\") pod \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\" (UID: \"1e013782-a27d-4cf6-bc7c-29a13bbd4137\") " Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.514637 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e013782-a27d-4cf6-bc7c-29a13bbd4137-config-volume" (OuterVolumeSpecName: "config-volume") pod "1e013782-a27d-4cf6-bc7c-29a13bbd4137" (UID: "1e013782-a27d-4cf6-bc7c-29a13bbd4137"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.533684 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e013782-a27d-4cf6-bc7c-29a13bbd4137-kube-api-access-pbzmh" (OuterVolumeSpecName: "kube-api-access-pbzmh") pod "1e013782-a27d-4cf6-bc7c-29a13bbd4137" (UID: "1e013782-a27d-4cf6-bc7c-29a13bbd4137"). InnerVolumeSpecName "kube-api-access-pbzmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.545173 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e013782-a27d-4cf6-bc7c-29a13bbd4137-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1e013782-a27d-4cf6-bc7c-29a13bbd4137" (UID: "1e013782-a27d-4cf6-bc7c-29a13bbd4137"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.614798 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e013782-a27d-4cf6-bc7c-29a13bbd4137-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.614861 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e013782-a27d-4cf6-bc7c-29a13bbd4137-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.614876 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbzmh\" (UniqueName: \"kubernetes.io/projected/1e013782-a27d-4cf6-bc7c-29a13bbd4137-kube-api-access-pbzmh\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.724466 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.811943 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" event={"ID":"1e013782-a27d-4cf6-bc7c-29a13bbd4137","Type":"ContainerDied","Data":"cd0e0b8af9898ed25e09ab5abd2e2899153dc5dc0b232d01e218e519a4a0d7f6"} Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.812006 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd0e0b8af9898ed25e09ab5abd2e2899153dc5dc0b232d01e218e519a4a0d7f6" Sep 30 14:45:04 crc kubenswrapper[4799]: I0930 14:45:04.812356 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49" Sep 30 14:45:05 crc kubenswrapper[4799]: I0930 14:45:05.133077 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.875245 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.876546 4799 generic.go:334] "Generic (PLEG): container finished" podID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerID="d0a099467cee71f3c37c55698919e290d84a71960fad48a65565f735a6ca095f" exitCode=0 Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.876728 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb82f13-0450-42a8-9c8a-0c174c534382","Type":"ContainerDied","Data":"d0a099467cee71f3c37c55698919e290d84a71960fad48a65565f735a6ca095f"} Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.887168 4799 generic.go:334] "Generic (PLEG): container finished" podID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerID="d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7" exitCode=0 Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.887233 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef6dcf7e-70a0-4759-afef-53e70b800ba0","Type":"ContainerDied","Data":"d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7"} Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.887268 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef6dcf7e-70a0-4759-afef-53e70b800ba0","Type":"ContainerDied","Data":"4a3f9ce55d0a4a85e0c47fcefbcac947f37982c69748a676af531e564ccf0152"} Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.887297 4799 scope.go:117] "RemoveContainer" containerID="d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7" Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.887493 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.945467 4799 scope.go:117] "RemoveContainer" containerID="4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3" Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994386 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-plugins-conf\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994524 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-erlang-cookie\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994561 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef6dcf7e-70a0-4759-afef-53e70b800ba0-pod-info\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994591 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz59q\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-kube-api-access-gz59q\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994623 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-config-data\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994705 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-confd\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994787 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-plugins\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994858 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-tls\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994884 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef6dcf7e-70a0-4759-afef-53e70b800ba0-erlang-cookie-secret\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994916 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:07 crc kubenswrapper[4799]: I0930 14:45:07.994970 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-server-conf\") pod \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\" (UID: \"ef6dcf7e-70a0-4759-afef-53e70b800ba0\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.001433 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.002929 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.013058 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.021434 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-kube-api-access-gz59q" (OuterVolumeSpecName: "kube-api-access-gz59q") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "kube-api-access-gz59q". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.023541 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.071563 4799 scope.go:117] "RemoveContainer" containerID="5090e4ee5ee0dc23dd5819cb106268ef2e915599e7e70f6901c27df62a123e6f" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.071727 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ef6dcf7e-70a0-4759-afef-53e70b800ba0-pod-info" (OuterVolumeSpecName: "pod-info") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.073285 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.077335 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef6dcf7e-70a0-4759-afef-53e70b800ba0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.079593 4799 scope.go:117] "RemoveContainer" containerID="d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7" Sep 30 14:45:08 crc kubenswrapper[4799]: E0930 14:45:08.080011 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7\": container with ID starting with d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7 not found: ID does not exist" containerID="d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.080049 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7"} err="failed to get container status \"d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7\": rpc error: code = NotFound desc = could not find container \"d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7\": container with ID starting with d0b7d01ad05e8c814c388b843e418ed0c3b00a578341c8e830173075096faed7 not found: ID does not exist" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.080076 4799 scope.go:117] "RemoveContainer" containerID="4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3" Sep 30 14:45:08 crc kubenswrapper[4799]: E0930 14:45:08.080299 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3\": container with ID starting with 4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3 not found: ID does not exist" containerID="4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.080324 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3"} err="failed to get container status \"4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3\": rpc error: code = NotFound desc = could not find container \"4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3\": container with ID starting with 4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3 not found: ID does not exist" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.101202 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.101247 4799 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef6dcf7e-70a0-4759-afef-53e70b800ba0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.101282 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.101295 4799 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.101307 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.101319 4799 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef6dcf7e-70a0-4759-afef-53e70b800ba0-pod-info\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.101330 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz59q\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-kube-api-access-gz59q\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.101340 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.112991 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-config-data" (OuterVolumeSpecName: "config-data") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.172663 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.176580 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.183462 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-server-conf" (OuterVolumeSpecName: "server-conf") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.203376 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.203691 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.204461 4799 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef6dcf7e-70a0-4759-afef-53e70b800ba0-server-conf\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.287788 4799 scope.go:117] "RemoveContainer" containerID="4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3" Sep 30 14:45:08 crc kubenswrapper[4799]: E0930 14:45:08.289583 4799 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3\": rpc error: code = NotFound desc = could not find container \"4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3\": container with ID starting with 4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3 not found: ID does not exist" containerID="4703c8944787fd6cb2a3214c20dd763017850590608c67cd28873fa3aa8618c3" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.289621 4799 scope.go:117] "RemoveContainer" containerID="d0a099467cee71f3c37c55698919e290d84a71960fad48a65565f735a6ca095f" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321040 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bb82f13-0450-42a8-9c8a-0c174c534382-erlang-cookie-secret\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321110 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bb82f13-0450-42a8-9c8a-0c174c534382-pod-info\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321272 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jb4r\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-kube-api-access-7jb4r\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321367 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-plugins\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321424 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-confd\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321447 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-plugins-conf\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321521 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-config-data\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321577 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-tls\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321697 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-erlang-cookie\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321740 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-server-conf\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.321769 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"1bb82f13-0450-42a8-9c8a-0c174c534382\" (UID: \"1bb82f13-0450-42a8-9c8a-0c174c534382\") " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.363172 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.380204 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.452388 4799 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-plugins-conf\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.452644 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.470706 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.471186 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.472287 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-config-data" (OuterVolumeSpecName: "config-data") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.473743 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bb82f13-0450-42a8-9c8a-0c174c534382-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.478014 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.478132 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-kube-api-access-7jb4r" (OuterVolumeSpecName: "kube-api-access-7jb4r") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "kube-api-access-7jb4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.547075 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-server-conf" (OuterVolumeSpecName: "server-conf") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.547274 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ef6dcf7e-70a0-4759-afef-53e70b800ba0" (UID: "ef6dcf7e-70a0-4759-afef-53e70b800ba0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.568979 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.569036 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.569050 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.569070 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.569090 4799 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bb82f13-0450-42a8-9c8a-0c174c534382-server-conf\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.569104 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef6dcf7e-70a0-4759-afef-53e70b800ba0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.569116 4799 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bb82f13-0450-42a8-9c8a-0c174c534382-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.569135 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jb4r\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-kube-api-access-7jb4r\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.605664 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1bb82f13-0450-42a8-9c8a-0c174c534382-pod-info" (OuterVolumeSpecName: "pod-info") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.617413 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1bb82f13-0450-42a8-9c8a-0c174c534382" (UID: "1bb82f13-0450-42a8-9c8a-0c174c534382"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.641981 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.674875 4799 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bb82f13-0450-42a8-9c8a-0c174c534382-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.674941 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.674957 4799 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bb82f13-0450-42a8-9c8a-0c174c534382-pod-info\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.815925 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.832003 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.864391 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:45:08 crc kubenswrapper[4799]: E0930 14:45:08.864915 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerName="rabbitmq" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.864938 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerName="rabbitmq" Sep 30 14:45:08 crc kubenswrapper[4799]: E0930 14:45:08.864968 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e013782-a27d-4cf6-bc7c-29a13bbd4137" containerName="collect-profiles" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.864978 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e013782-a27d-4cf6-bc7c-29a13bbd4137" containerName="collect-profiles" Sep 30 14:45:08 crc kubenswrapper[4799]: E0930 14:45:08.865010 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerName="setup-container" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.865017 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerName="setup-container" Sep 30 14:45:08 crc kubenswrapper[4799]: E0930 14:45:08.865027 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerName="rabbitmq" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.865033 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerName="rabbitmq" Sep 30 14:45:08 crc kubenswrapper[4799]: E0930 14:45:08.865052 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerName="setup-container" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.865060 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerName="setup-container" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.865280 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" containerName="rabbitmq" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.865300 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" containerName="rabbitmq" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.865316 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e013782-a27d-4cf6-bc7c-29a13bbd4137" containerName="collect-profiles" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.866781 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.870118 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.870159 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.870301 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.870364 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.870452 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.870121 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.870520 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g4bn4" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.904619 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.927711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb82f13-0450-42a8-9c8a-0c174c534382","Type":"ContainerDied","Data":"0e93bb0e81164ea28f99c0b1bd326238744ae6c6f285ce255c65c499dd713c96"} Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.927777 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.927790 4799 scope.go:117] "RemoveContainer" containerID="0817b5a4428f559089509977d4a81d1d3084c69e64b252f2427d59eae955efa7" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.986130 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmzqz\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-kube-api-access-cmzqz\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.986203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.986295 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.986372 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f6f049b3-7415-4acc-829a-659677059137-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.986436 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.986466 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.986564 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.991497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.992149 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f6f049b3-7415-4acc-829a-659677059137-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.992203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:08 crc kubenswrapper[4799]: I0930 14:45:08.992270 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.011741 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.034884 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.055935 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.058406 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.061853 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.071169 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.071240 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.071189 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.071485 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.071558 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.071580 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-qd87k" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.080319 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096323 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096392 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096533 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096629 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096702 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f6f049b3-7415-4acc-829a-659677059137-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096739 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096787 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096852 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmzqz\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-kube-api-access-cmzqz\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096893 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096925 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.096988 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f6f049b3-7415-4acc-829a-659677059137-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.097878 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.098129 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.098179 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.099395 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.100197 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.102012 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f6f049b3-7415-4acc-829a-659677059137-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.116167 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f6f049b3-7415-4acc-829a-659677059137-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.127243 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.128209 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.129070 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f6f049b3-7415-4acc-829a-659677059137-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.164551 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmzqz\" (UniqueName: \"kubernetes.io/projected/f6f049b3-7415-4acc-829a-659677059137-kube-api-access-cmzqz\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.199680 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200073 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200109 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200162 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h5wg\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-kube-api-access-8h5wg\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200202 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200249 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200297 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200355 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200422 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200467 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.200518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.206461 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f6f049b3-7415-4acc-829a-659677059137\") " pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.302994 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303155 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303198 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303231 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303279 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h5wg\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-kube-api-access-8h5wg\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303323 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303365 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303407 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.303517 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.304240 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.304702 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.305366 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.306592 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.309400 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.310213 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-config-data\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.310414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.310562 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.325508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.327413 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.333371 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h5wg\" (UniqueName: \"kubernetes.io/projected/1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2-kube-api-access-8h5wg\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.364952 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2\") " pod="openstack/rabbitmq-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.495293 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:09 crc kubenswrapper[4799]: I0930 14:45:09.584470 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Sep 30 14:45:10 crc kubenswrapper[4799]: I0930 14:45:10.242213 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Sep 30 14:45:10 crc kubenswrapper[4799]: I0930 14:45:10.428473 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Sep 30 14:45:10 crc kubenswrapper[4799]: I0930 14:45:10.642707 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bb82f13-0450-42a8-9c8a-0c174c534382" path="/var/lib/kubelet/pods/1bb82f13-0450-42a8-9c8a-0c174c534382/volumes" Sep 30 14:45:10 crc kubenswrapper[4799]: I0930 14:45:10.646595 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef6dcf7e-70a0-4759-afef-53e70b800ba0" path="/var/lib/kubelet/pods/ef6dcf7e-70a0-4759-afef-53e70b800ba0/volumes" Sep 30 14:45:10 crc kubenswrapper[4799]: I0930 14:45:10.967793 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2","Type":"ContainerStarted","Data":"3039f575a58d2c5c3176af0882334969c6332d4191db75fbf8b2c85c0573c6cd"} Sep 30 14:45:10 crc kubenswrapper[4799]: I0930 14:45:10.969437 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f6f049b3-7415-4acc-829a-659677059137","Type":"ContainerStarted","Data":"4afc4ccbada1524f2973351e1de6e2f7bead262e80aa662756da11f94e4566fb"} Sep 30 14:45:11 crc kubenswrapper[4799]: I0930 14:45:11.983295 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2","Type":"ContainerStarted","Data":"a42cc36fe04db8321f37605db23c2785ce1efd063c0416f0e5f3c1324267e720"} Sep 30 14:45:11 crc kubenswrapper[4799]: I0930 14:45:11.986241 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f6f049b3-7415-4acc-829a-659677059137","Type":"ContainerStarted","Data":"04bdf5c0499240540380a7c77cde5cf0f74a056d2cb58f28409cec3573ea2ed4"} Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.335928 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-gkf96"] Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.338590 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.342192 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.351921 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-gkf96"] Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.448804 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd56f\" (UniqueName: \"kubernetes.io/projected/6fd1b446-a308-442d-8879-a6fca8c66491-kube-api-access-xd56f\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.448936 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.448959 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.448998 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.449042 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-svc\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.449071 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.449280 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-config\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.551698 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd56f\" (UniqueName: \"kubernetes.io/projected/6fd1b446-a308-442d-8879-a6fca8c66491-kube-api-access-xd56f\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.551854 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.551898 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.551949 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.552035 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-svc\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.552082 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.552137 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-config\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.553459 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-config\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.553617 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.553779 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-svc\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.553798 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.553872 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.553897 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.576886 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd56f\" (UniqueName: \"kubernetes.io/projected/6fd1b446-a308-442d-8879-a6fca8c66491-kube-api-access-xd56f\") pod \"dnsmasq-dns-67b789f86c-gkf96\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:13 crc kubenswrapper[4799]: I0930 14:45:13.675956 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:14 crc kubenswrapper[4799]: W0930 14:45:14.856544 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fd1b446_a308_442d_8879_a6fca8c66491.slice/crio-b6c6d1e03b63fe2cffa8a3d5ed77bff325ed0cf93f77e44f9480ffd13115f60b WatchSource:0}: Error finding container b6c6d1e03b63fe2cffa8a3d5ed77bff325ed0cf93f77e44f9480ffd13115f60b: Status 404 returned error can't find the container with id b6c6d1e03b63fe2cffa8a3d5ed77bff325ed0cf93f77e44f9480ffd13115f60b Sep 30 14:45:14 crc kubenswrapper[4799]: I0930 14:45:14.904556 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-gkf96"] Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.021925 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" event={"ID":"6fd1b446-a308-442d-8879-a6fca8c66491","Type":"ContainerStarted","Data":"b6c6d1e03b63fe2cffa8a3d5ed77bff325ed0cf93f77e44f9480ffd13115f60b"} Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.026797 4799 generic.go:334] "Generic (PLEG): container finished" podID="d15890bb-d088-4126-8382-b5d8e423b608" containerID="15813eeea4cf4efe8492111998588d8903edae8edcbbf0a1ab40735200a85876" exitCode=137 Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.027098 4799 generic.go:334] "Generic (PLEG): container finished" podID="d15890bb-d088-4126-8382-b5d8e423b608" containerID="8923b064d4910eddccd36258c0091f291d87f5d2b5ba8a2a2855d3bafc5f9665" exitCode=137 Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.027070 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerDied","Data":"15813eeea4cf4efe8492111998588d8903edae8edcbbf0a1ab40735200a85876"} Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.027305 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerDied","Data":"8923b064d4910eddccd36258c0091f291d87f5d2b5ba8a2a2855d3bafc5f9665"} Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.027398 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f99689854-glm7t" event={"ID":"d15890bb-d088-4126-8382-b5d8e423b608","Type":"ContainerDied","Data":"65cfd9ee288b6b3d104f178a313a71390fc9ae06d91f2a0f13adf2190c01ea70"} Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.027481 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65cfd9ee288b6b3d104f178a313a71390fc9ae06d91f2a0f13adf2190c01ea70" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.027380 4799 scope.go:117] "RemoveContainer" containerID="e36dbd20d5d3cdf7ece15c8b2611df5884ae476a6712a79abbb5e85a1de0277d" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.154252 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.195788 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-config-data\") pod \"d15890bb-d088-4126-8382-b5d8e423b608\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.195879 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-tls-certs\") pod \"d15890bb-d088-4126-8382-b5d8e423b608\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.195939 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-scripts\") pod \"d15890bb-d088-4126-8382-b5d8e423b608\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.196035 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15890bb-d088-4126-8382-b5d8e423b608-logs\") pod \"d15890bb-d088-4126-8382-b5d8e423b608\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.196120 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8ggr\" (UniqueName: \"kubernetes.io/projected/d15890bb-d088-4126-8382-b5d8e423b608-kube-api-access-b8ggr\") pod \"d15890bb-d088-4126-8382-b5d8e423b608\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.196174 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-combined-ca-bundle\") pod \"d15890bb-d088-4126-8382-b5d8e423b608\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.196232 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-secret-key\") pod \"d15890bb-d088-4126-8382-b5d8e423b608\" (UID: \"d15890bb-d088-4126-8382-b5d8e423b608\") " Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.224968 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d15890bb-d088-4126-8382-b5d8e423b608" (UID: "d15890bb-d088-4126-8382-b5d8e423b608"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.225397 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d15890bb-d088-4126-8382-b5d8e423b608-logs" (OuterVolumeSpecName: "logs") pod "d15890bb-d088-4126-8382-b5d8e423b608" (UID: "d15890bb-d088-4126-8382-b5d8e423b608"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.226088 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d15890bb-d088-4126-8382-b5d8e423b608-kube-api-access-b8ggr" (OuterVolumeSpecName: "kube-api-access-b8ggr") pod "d15890bb-d088-4126-8382-b5d8e423b608" (UID: "d15890bb-d088-4126-8382-b5d8e423b608"). InnerVolumeSpecName "kube-api-access-b8ggr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.298504 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.298541 4799 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15890bb-d088-4126-8382-b5d8e423b608-logs\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.298552 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8ggr\" (UniqueName: \"kubernetes.io/projected/d15890bb-d088-4126-8382-b5d8e423b608-kube-api-access-b8ggr\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.342875 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-config-data" (OuterVolumeSpecName: "config-data") pod "d15890bb-d088-4126-8382-b5d8e423b608" (UID: "d15890bb-d088-4126-8382-b5d8e423b608"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.345098 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d15890bb-d088-4126-8382-b5d8e423b608" (UID: "d15890bb-d088-4126-8382-b5d8e423b608"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.368863 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-scripts" (OuterVolumeSpecName: "scripts") pod "d15890bb-d088-4126-8382-b5d8e423b608" (UID: "d15890bb-d088-4126-8382-b5d8e423b608"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.389786 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "d15890bb-d088-4126-8382-b5d8e423b608" (UID: "d15890bb-d088-4126-8382-b5d8e423b608"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.400434 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.400471 4799 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.400484 4799 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d15890bb-d088-4126-8382-b5d8e423b608-scripts\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:15 crc kubenswrapper[4799]: I0930 14:45:15.400503 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15890bb-d088-4126-8382-b5d8e423b608-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:16 crc kubenswrapper[4799]: I0930 14:45:16.043055 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f99689854-glm7t" Sep 30 14:45:16 crc kubenswrapper[4799]: I0930 14:45:16.047036 4799 generic.go:334] "Generic (PLEG): container finished" podID="6fd1b446-a308-442d-8879-a6fca8c66491" containerID="326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0" exitCode=0 Sep 30 14:45:16 crc kubenswrapper[4799]: I0930 14:45:16.047100 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" event={"ID":"6fd1b446-a308-442d-8879-a6fca8c66491","Type":"ContainerDied","Data":"326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0"} Sep 30 14:45:16 crc kubenswrapper[4799]: I0930 14:45:16.290613 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f99689854-glm7t"] Sep 30 14:45:16 crc kubenswrapper[4799]: I0930 14:45:16.300328 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f99689854-glm7t"] Sep 30 14:45:16 crc kubenswrapper[4799]: I0930 14:45:16.516021 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d15890bb-d088-4126-8382-b5d8e423b608" path="/var/lib/kubelet/pods/d15890bb-d088-4126-8382-b5d8e423b608/volumes" Sep 30 14:45:17 crc kubenswrapper[4799]: I0930 14:45:17.062807 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" event={"ID":"6fd1b446-a308-442d-8879-a6fca8c66491","Type":"ContainerStarted","Data":"d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c"} Sep 30 14:45:17 crc kubenswrapper[4799]: I0930 14:45:17.064379 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:17 crc kubenswrapper[4799]: I0930 14:45:17.096460 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" podStartSLOduration=4.096433331 podStartE2EDuration="4.096433331s" podCreationTimestamp="2025-09-30 14:45:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:45:17.089381885 +0000 UTC m=+1539.172982322" watchObservedRunningTime="2025-09-30 14:45:17.096433331 +0000 UTC m=+1539.180033758" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.677971 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.750327 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-xwzw6"] Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.750652 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" podUID="9f78baee-06bd-4654-bc64-c9919f4dd420" containerName="dnsmasq-dns" containerID="cri-o://8b70e418509f54fc6828cf833c4b7035756b4d2c3f7fde7183a6a3c67a912968" gracePeriod=10 Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.996449 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-s2gr4"] Sep 30 14:45:23 crc kubenswrapper[4799]: E0930 14:45:23.997068 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon-log" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997091 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon-log" Sep 30 14:45:23 crc kubenswrapper[4799]: E0930 14:45:23.997107 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997113 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: E0930 14:45:23.997133 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997138 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: E0930 14:45:23.997150 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997157 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997450 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997480 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997497 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997510 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997531 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon-log" Sep 30 14:45:23 crc kubenswrapper[4799]: E0930 14:45:23.997762 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997774 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: E0930 14:45:23.997783 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997789 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.997991 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15890bb-d088-4126-8382-b5d8e423b608" containerName="horizon" Sep 30 14:45:23 crc kubenswrapper[4799]: I0930 14:45:23.998858 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.035617 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-s2gr4"] Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.104134 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.104256 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.104355 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbls4\" (UniqueName: \"kubernetes.io/projected/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-kube-api-access-sbls4\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.104395 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.104425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.104455 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.104547 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-config\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.148835 4799 generic.go:334] "Generic (PLEG): container finished" podID="9f78baee-06bd-4654-bc64-c9919f4dd420" containerID="8b70e418509f54fc6828cf833c4b7035756b4d2c3f7fde7183a6a3c67a912968" exitCode=0 Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.148907 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" event={"ID":"9f78baee-06bd-4654-bc64-c9919f4dd420","Type":"ContainerDied","Data":"8b70e418509f54fc6828cf833c4b7035756b4d2c3f7fde7183a6a3c67a912968"} Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.206375 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.206458 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.206486 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbls4\" (UniqueName: \"kubernetes.io/projected/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-kube-api-access-sbls4\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.206511 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.206535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.206556 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.206633 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-config\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.207401 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.207948 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-config\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.208346 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.208392 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.208690 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.209363 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.244475 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbls4\" (UniqueName: \"kubernetes.io/projected/9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd-kube-api-access-sbls4\") pod \"dnsmasq-dns-79dc84bdb7-s2gr4\" (UID: \"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd\") " pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.324300 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.479358 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.616716 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-swift-storage-0\") pod \"9f78baee-06bd-4654-bc64-c9919f4dd420\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.617220 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-sb\") pod \"9f78baee-06bd-4654-bc64-c9919f4dd420\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.617320 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-nb\") pod \"9f78baee-06bd-4654-bc64-c9919f4dd420\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.617428 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-svc\") pod \"9f78baee-06bd-4654-bc64-c9919f4dd420\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.617525 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbldf\" (UniqueName: \"kubernetes.io/projected/9f78baee-06bd-4654-bc64-c9919f4dd420-kube-api-access-hbldf\") pod \"9f78baee-06bd-4654-bc64-c9919f4dd420\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.617857 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-config\") pod \"9f78baee-06bd-4654-bc64-c9919f4dd420\" (UID: \"9f78baee-06bd-4654-bc64-c9919f4dd420\") " Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.631783 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f78baee-06bd-4654-bc64-c9919f4dd420-kube-api-access-hbldf" (OuterVolumeSpecName: "kube-api-access-hbldf") pod "9f78baee-06bd-4654-bc64-c9919f4dd420" (UID: "9f78baee-06bd-4654-bc64-c9919f4dd420"). InnerVolumeSpecName "kube-api-access-hbldf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.722172 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbldf\" (UniqueName: \"kubernetes.io/projected/9f78baee-06bd-4654-bc64-c9919f4dd420-kube-api-access-hbldf\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.758464 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f78baee-06bd-4654-bc64-c9919f4dd420" (UID: "9f78baee-06bd-4654-bc64-c9919f4dd420"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.805901 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9f78baee-06bd-4654-bc64-c9919f4dd420" (UID: "9f78baee-06bd-4654-bc64-c9919f4dd420"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.830764 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.830809 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.830938 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9f78baee-06bd-4654-bc64-c9919f4dd420" (UID: "9f78baee-06bd-4654-bc64-c9919f4dd420"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.861302 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-config" (OuterVolumeSpecName: "config") pod "9f78baee-06bd-4654-bc64-c9919f4dd420" (UID: "9f78baee-06bd-4654-bc64-c9919f4dd420"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.876824 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9f78baee-06bd-4654-bc64-c9919f4dd420" (UID: "9f78baee-06bd-4654-bc64-c9919f4dd420"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.937201 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.937247 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.937260 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f78baee-06bd-4654-bc64-c9919f4dd420-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:24 crc kubenswrapper[4799]: I0930 14:45:24.999100 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-s2gr4"] Sep 30 14:45:25 crc kubenswrapper[4799]: I0930 14:45:25.165674 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" event={"ID":"9f78baee-06bd-4654-bc64-c9919f4dd420","Type":"ContainerDied","Data":"d2c142405e8ef5a79c0b5339069481931d2ae923d5541b8088c2980730345262"} Sep 30 14:45:25 crc kubenswrapper[4799]: I0930 14:45:25.166180 4799 scope.go:117] "RemoveContainer" containerID="8b70e418509f54fc6828cf833c4b7035756b4d2c3f7fde7183a6a3c67a912968" Sep 30 14:45:25 crc kubenswrapper[4799]: I0930 14:45:25.165730 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-xwzw6" Sep 30 14:45:25 crc kubenswrapper[4799]: I0930 14:45:25.167597 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" event={"ID":"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd","Type":"ContainerStarted","Data":"6efcdb83c8a79e7685e823f67ab3dbc223ab5d7a462dea1b268153d1c0306e29"} Sep 30 14:45:25 crc kubenswrapper[4799]: I0930 14:45:25.191989 4799 scope.go:117] "RemoveContainer" containerID="2826ad43b3d354df327ca59f205e8df8bc37f64b5ee8ea1c9f31ab4c93d8cc65" Sep 30 14:45:25 crc kubenswrapper[4799]: I0930 14:45:25.211258 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-xwzw6"] Sep 30 14:45:25 crc kubenswrapper[4799]: I0930 14:45:25.224842 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-xwzw6"] Sep 30 14:45:26 crc kubenswrapper[4799]: I0930 14:45:26.180758 4799 generic.go:334] "Generic (PLEG): container finished" podID="9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd" containerID="86332d1df4852dfaa3f38fa075679cc407608bf47b66af2bb6d48ad7f829f6f2" exitCode=0 Sep 30 14:45:26 crc kubenswrapper[4799]: I0930 14:45:26.180848 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" event={"ID":"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd","Type":"ContainerDied","Data":"86332d1df4852dfaa3f38fa075679cc407608bf47b66af2bb6d48ad7f829f6f2"} Sep 30 14:45:26 crc kubenswrapper[4799]: I0930 14:45:26.516696 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f78baee-06bd-4654-bc64-c9919f4dd420" path="/var/lib/kubelet/pods/9f78baee-06bd-4654-bc64-c9919f4dd420/volumes" Sep 30 14:45:27 crc kubenswrapper[4799]: I0930 14:45:27.195096 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" event={"ID":"9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd","Type":"ContainerStarted","Data":"0da2a29cc4ded18fdc3f9419170082126619e9d5fb3b1160d58506bc955f7df6"} Sep 30 14:45:27 crc kubenswrapper[4799]: I0930 14:45:27.195300 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:27 crc kubenswrapper[4799]: I0930 14:45:27.220314 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" podStartSLOduration=4.22028907 podStartE2EDuration="4.22028907s" podCreationTimestamp="2025-09-30 14:45:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:45:27.217361785 +0000 UTC m=+1549.300962212" watchObservedRunningTime="2025-09-30 14:45:27.22028907 +0000 UTC m=+1549.303889497" Sep 30 14:45:29 crc kubenswrapper[4799]: I0930 14:45:29.651893 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:45:29 crc kubenswrapper[4799]: I0930 14:45:29.652282 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:45:29 crc kubenswrapper[4799]: I0930 14:45:29.652348 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:45:29 crc kubenswrapper[4799]: I0930 14:45:29.653364 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:45:29 crc kubenswrapper[4799]: I0930 14:45:29.653435 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" gracePeriod=600 Sep 30 14:45:29 crc kubenswrapper[4799]: E0930 14:45:29.811272 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:45:30 crc kubenswrapper[4799]: I0930 14:45:30.243888 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" exitCode=0 Sep 30 14:45:30 crc kubenswrapper[4799]: I0930 14:45:30.244148 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1"} Sep 30 14:45:30 crc kubenswrapper[4799]: I0930 14:45:30.244229 4799 scope.go:117] "RemoveContainer" containerID="bb9ad6936dc7236258e4efd5f74222bbe45a21ab72fed8d7b84c8d34eafe641b" Sep 30 14:45:30 crc kubenswrapper[4799]: I0930 14:45:30.245194 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:45:30 crc kubenswrapper[4799]: E0930 14:45:30.245484 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:45:34 crc kubenswrapper[4799]: I0930 14:45:34.327712 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79dc84bdb7-s2gr4" Sep 30 14:45:34 crc kubenswrapper[4799]: I0930 14:45:34.408165 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-gkf96"] Sep 30 14:45:34 crc kubenswrapper[4799]: I0930 14:45:34.408937 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" podUID="6fd1b446-a308-442d-8879-a6fca8c66491" containerName="dnsmasq-dns" containerID="cri-o://d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c" gracePeriod=10 Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:34.999811 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.062413 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-openstack-edpm-ipam\") pod \"6fd1b446-a308-442d-8879-a6fca8c66491\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.062468 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-swift-storage-0\") pod \"6fd1b446-a308-442d-8879-a6fca8c66491\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.062515 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-sb\") pod \"6fd1b446-a308-442d-8879-a6fca8c66491\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.062549 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-config\") pod \"6fd1b446-a308-442d-8879-a6fca8c66491\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.062644 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-svc\") pod \"6fd1b446-a308-442d-8879-a6fca8c66491\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.062801 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd56f\" (UniqueName: \"kubernetes.io/projected/6fd1b446-a308-442d-8879-a6fca8c66491-kube-api-access-xd56f\") pod \"6fd1b446-a308-442d-8879-a6fca8c66491\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.062898 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-nb\") pod \"6fd1b446-a308-442d-8879-a6fca8c66491\" (UID: \"6fd1b446-a308-442d-8879-a6fca8c66491\") " Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.078964 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fd1b446-a308-442d-8879-a6fca8c66491-kube-api-access-xd56f" (OuterVolumeSpecName: "kube-api-access-xd56f") pod "6fd1b446-a308-442d-8879-a6fca8c66491" (UID: "6fd1b446-a308-442d-8879-a6fca8c66491"). InnerVolumeSpecName "kube-api-access-xd56f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.154120 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6fd1b446-a308-442d-8879-a6fca8c66491" (UID: "6fd1b446-a308-442d-8879-a6fca8c66491"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.160338 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6fd1b446-a308-442d-8879-a6fca8c66491" (UID: "6fd1b446-a308-442d-8879-a6fca8c66491"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.161295 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-config" (OuterVolumeSpecName: "config") pod "6fd1b446-a308-442d-8879-a6fca8c66491" (UID: "6fd1b446-a308-442d-8879-a6fca8c66491"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.165942 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd56f\" (UniqueName: \"kubernetes.io/projected/6fd1b446-a308-442d-8879-a6fca8c66491-kube-api-access-xd56f\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.166112 4799 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.166181 4799 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-config\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.166326 4799 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-dns-svc\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.170645 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "6fd1b446-a308-442d-8879-a6fca8c66491" (UID: "6fd1b446-a308-442d-8879-a6fca8c66491"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.191027 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6fd1b446-a308-442d-8879-a6fca8c66491" (UID: "6fd1b446-a308-442d-8879-a6fca8c66491"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.203002 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6fd1b446-a308-442d-8879-a6fca8c66491" (UID: "6fd1b446-a308-442d-8879-a6fca8c66491"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.268702 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.268994 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.269007 4799 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fd1b446-a308-442d-8879-a6fca8c66491-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.306870 4799 generic.go:334] "Generic (PLEG): container finished" podID="6fd1b446-a308-442d-8879-a6fca8c66491" containerID="d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c" exitCode=0 Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.307051 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" event={"ID":"6fd1b446-a308-442d-8879-a6fca8c66491","Type":"ContainerDied","Data":"d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c"} Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.307254 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" event={"ID":"6fd1b446-a308-442d-8879-a6fca8c66491","Type":"ContainerDied","Data":"b6c6d1e03b63fe2cffa8a3d5ed77bff325ed0cf93f77e44f9480ffd13115f60b"} Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.307147 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-gkf96" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.307324 4799 scope.go:117] "RemoveContainer" containerID="d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.344091 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-gkf96"] Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.346598 4799 scope.go:117] "RemoveContainer" containerID="326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.353254 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-gkf96"] Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.369764 4799 scope.go:117] "RemoveContainer" containerID="d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c" Sep 30 14:45:35 crc kubenswrapper[4799]: E0930 14:45:35.370464 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c\": container with ID starting with d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c not found: ID does not exist" containerID="d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.370539 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c"} err="failed to get container status \"d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c\": rpc error: code = NotFound desc = could not find container \"d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c\": container with ID starting with d79ad1a1259a1b44b1f678d37a86a26bb7be013748906a42469ac626a015775c not found: ID does not exist" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.370596 4799 scope.go:117] "RemoveContainer" containerID="326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0" Sep 30 14:45:35 crc kubenswrapper[4799]: E0930 14:45:35.371843 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0\": container with ID starting with 326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0 not found: ID does not exist" containerID="326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0" Sep 30 14:45:35 crc kubenswrapper[4799]: I0930 14:45:35.371871 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0"} err="failed to get container status \"326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0\": rpc error: code = NotFound desc = could not find container \"326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0\": container with ID starting with 326726815cbfcca3ea1e2f077700e6cc62e92f7b6ff32175310098bded7d79f0 not found: ID does not exist" Sep 30 14:45:36 crc kubenswrapper[4799]: I0930 14:45:36.516278 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fd1b446-a308-442d-8879-a6fca8c66491" path="/var/lib/kubelet/pods/6fd1b446-a308-442d-8879-a6fca8c66491/volumes" Sep 30 14:45:41 crc kubenswrapper[4799]: E0930 14:45:41.558393 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6f049b3_7415_4acc_829a_659677059137.slice/crio-conmon-04bdf5c0499240540380a7c77cde5cf0f74a056d2cb58f28409cec3573ea2ed4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6f049b3_7415_4acc_829a_659677059137.slice/crio-04bdf5c0499240540380a7c77cde5cf0f74a056d2cb58f28409cec3573ea2ed4.scope\": RecentStats: unable to find data in memory cache]" Sep 30 14:45:42 crc kubenswrapper[4799]: I0930 14:45:42.393494 4799 generic.go:334] "Generic (PLEG): container finished" podID="1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2" containerID="a42cc36fe04db8321f37605db23c2785ce1efd063c0416f0e5f3c1324267e720" exitCode=0 Sep 30 14:45:42 crc kubenswrapper[4799]: I0930 14:45:42.393593 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2","Type":"ContainerDied","Data":"a42cc36fe04db8321f37605db23c2785ce1efd063c0416f0e5f3c1324267e720"} Sep 30 14:45:42 crc kubenswrapper[4799]: I0930 14:45:42.396111 4799 generic.go:334] "Generic (PLEG): container finished" podID="f6f049b3-7415-4acc-829a-659677059137" containerID="04bdf5c0499240540380a7c77cde5cf0f74a056d2cb58f28409cec3573ea2ed4" exitCode=0 Sep 30 14:45:42 crc kubenswrapper[4799]: I0930 14:45:42.396144 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f6f049b3-7415-4acc-829a-659677059137","Type":"ContainerDied","Data":"04bdf5c0499240540380a7c77cde5cf0f74a056d2cb58f28409cec3573ea2ed4"} Sep 30 14:45:43 crc kubenswrapper[4799]: I0930 14:45:43.408711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2","Type":"ContainerStarted","Data":"7375d2f01f596f798fc1b5a2739261fd1d66b2071865a3a13e1224917e49f967"} Sep 30 14:45:43 crc kubenswrapper[4799]: I0930 14:45:43.409466 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Sep 30 14:45:43 crc kubenswrapper[4799]: I0930 14:45:43.413184 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f6f049b3-7415-4acc-829a-659677059137","Type":"ContainerStarted","Data":"43484c28fbf3bacf1d6a2c58fdd2519feed9b12c9f3b426ce98c488008d9d810"} Sep 30 14:45:43 crc kubenswrapper[4799]: I0930 14:45:43.413403 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:43 crc kubenswrapper[4799]: I0930 14:45:43.442505 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.442472768 podStartE2EDuration="35.442472768s" podCreationTimestamp="2025-09-30 14:45:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:45:43.435865615 +0000 UTC m=+1565.519466042" watchObservedRunningTime="2025-09-30 14:45:43.442472768 +0000 UTC m=+1565.526073195" Sep 30 14:45:43 crc kubenswrapper[4799]: I0930 14:45:43.470051 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.470027114 podStartE2EDuration="35.470027114s" podCreationTimestamp="2025-09-30 14:45:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 14:45:43.463367479 +0000 UTC m=+1565.546967926" watchObservedRunningTime="2025-09-30 14:45:43.470027114 +0000 UTC m=+1565.553627541" Sep 30 14:45:44 crc kubenswrapper[4799]: I0930 14:45:44.504761 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:45:44 crc kubenswrapper[4799]: E0930 14:45:44.505050 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.747228 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7"] Sep 30 14:45:52 crc kubenswrapper[4799]: E0930 14:45:52.749646 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd1b446-a308-442d-8879-a6fca8c66491" containerName="dnsmasq-dns" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.749816 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd1b446-a308-442d-8879-a6fca8c66491" containerName="dnsmasq-dns" Sep 30 14:45:52 crc kubenswrapper[4799]: E0930 14:45:52.749907 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f78baee-06bd-4654-bc64-c9919f4dd420" containerName="init" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.749984 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f78baee-06bd-4654-bc64-c9919f4dd420" containerName="init" Sep 30 14:45:52 crc kubenswrapper[4799]: E0930 14:45:52.750051 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd1b446-a308-442d-8879-a6fca8c66491" containerName="init" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.750117 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd1b446-a308-442d-8879-a6fca8c66491" containerName="init" Sep 30 14:45:52 crc kubenswrapper[4799]: E0930 14:45:52.750204 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f78baee-06bd-4654-bc64-c9919f4dd420" containerName="dnsmasq-dns" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.750278 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f78baee-06bd-4654-bc64-c9919f4dd420" containerName="dnsmasq-dns" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.751951 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fd1b446-a308-442d-8879-a6fca8c66491" containerName="dnsmasq-dns" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.752085 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f78baee-06bd-4654-bc64-c9919f4dd420" containerName="dnsmasq-dns" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.753103 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.755686 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.756233 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.756503 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.756810 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.771229 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7"] Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.848206 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.848289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.848372 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksw5x\" (UniqueName: \"kubernetes.io/projected/2fdc7b08-b5e8-45cb-97d9-d44658909edf-kube-api-access-ksw5x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.848398 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.949856 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksw5x\" (UniqueName: \"kubernetes.io/projected/2fdc7b08-b5e8-45cb-97d9-d44658909edf-kube-api-access-ksw5x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.949921 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.949997 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.950045 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.956469 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.956841 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.957208 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:52 crc kubenswrapper[4799]: I0930 14:45:52.973757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksw5x\" (UniqueName: \"kubernetes.io/projected/2fdc7b08-b5e8-45cb-97d9-d44658909edf-kube-api-access-ksw5x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:53 crc kubenswrapper[4799]: I0930 14:45:53.073758 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:45:54 crc kubenswrapper[4799]: I0930 14:45:54.017795 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7"] Sep 30 14:45:54 crc kubenswrapper[4799]: I0930 14:45:54.524949 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" event={"ID":"2fdc7b08-b5e8-45cb-97d9-d44658909edf","Type":"ContainerStarted","Data":"5167e632613ea6e9935e02cb1f76fbc9fd95ba43065afa177434c9005e783902"} Sep 30 14:45:56 crc kubenswrapper[4799]: I0930 14:45:56.504591 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:45:56 crc kubenswrapper[4799]: E0930 14:45:56.504986 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:45:59 crc kubenswrapper[4799]: I0930 14:45:59.499560 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Sep 30 14:45:59 crc kubenswrapper[4799]: I0930 14:45:59.589880 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Sep 30 14:46:08 crc kubenswrapper[4799]: I0930 14:46:08.664795 4799 scope.go:117] "RemoveContainer" containerID="f5239377f75f41afe5178e8987a478a2f4461fff97d01826fa1359713ddf2fce" Sep 30 14:46:09 crc kubenswrapper[4799]: I0930 14:46:09.507335 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:46:09 crc kubenswrapper[4799]: E0930 14:46:09.507963 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:46:09 crc kubenswrapper[4799]: I0930 14:46:09.756552 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" event={"ID":"2fdc7b08-b5e8-45cb-97d9-d44658909edf","Type":"ContainerStarted","Data":"18c9de60f41fc05af7c8d43661b4e663004f184af486e0e6562b3f1f6f3c98de"} Sep 30 14:46:09 crc kubenswrapper[4799]: I0930 14:46:09.785045 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" podStartSLOduration=3.261448384 podStartE2EDuration="17.785013235s" podCreationTimestamp="2025-09-30 14:45:52 +0000 UTC" firstStartedPulling="2025-09-30 14:45:54.040673055 +0000 UTC m=+1576.124273482" lastFinishedPulling="2025-09-30 14:46:08.564237906 +0000 UTC m=+1590.647838333" observedRunningTime="2025-09-30 14:46:09.781371228 +0000 UTC m=+1591.864971665" watchObservedRunningTime="2025-09-30 14:46:09.785013235 +0000 UTC m=+1591.868613662" Sep 30 14:46:20 crc kubenswrapper[4799]: I0930 14:46:20.883837 4799 generic.go:334] "Generic (PLEG): container finished" podID="2fdc7b08-b5e8-45cb-97d9-d44658909edf" containerID="18c9de60f41fc05af7c8d43661b4e663004f184af486e0e6562b3f1f6f3c98de" exitCode=0 Sep 30 14:46:20 crc kubenswrapper[4799]: I0930 14:46:20.883924 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" event={"ID":"2fdc7b08-b5e8-45cb-97d9-d44658909edf","Type":"ContainerDied","Data":"18c9de60f41fc05af7c8d43661b4e663004f184af486e0e6562b3f1f6f3c98de"} Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.321689 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.326117 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-inventory\") pod \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.326296 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-ssh-key\") pod \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.326439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-repo-setup-combined-ca-bundle\") pod \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.326548 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksw5x\" (UniqueName: \"kubernetes.io/projected/2fdc7b08-b5e8-45cb-97d9-d44658909edf-kube-api-access-ksw5x\") pod \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\" (UID: \"2fdc7b08-b5e8-45cb-97d9-d44658909edf\") " Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.333572 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2fdc7b08-b5e8-45cb-97d9-d44658909edf" (UID: "2fdc7b08-b5e8-45cb-97d9-d44658909edf"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.336859 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fdc7b08-b5e8-45cb-97d9-d44658909edf-kube-api-access-ksw5x" (OuterVolumeSpecName: "kube-api-access-ksw5x") pod "2fdc7b08-b5e8-45cb-97d9-d44658909edf" (UID: "2fdc7b08-b5e8-45cb-97d9-d44658909edf"). InnerVolumeSpecName "kube-api-access-ksw5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.368023 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-inventory" (OuterVolumeSpecName: "inventory") pod "2fdc7b08-b5e8-45cb-97d9-d44658909edf" (UID: "2fdc7b08-b5e8-45cb-97d9-d44658909edf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.374738 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2fdc7b08-b5e8-45cb-97d9-d44658909edf" (UID: "2fdc7b08-b5e8-45cb-97d9-d44658909edf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.429528 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.429568 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.429580 4799 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fdc7b08-b5e8-45cb-97d9-d44658909edf-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.429592 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksw5x\" (UniqueName: \"kubernetes.io/projected/2fdc7b08-b5e8-45cb-97d9-d44658909edf-kube-api-access-ksw5x\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:22 crc kubenswrapper[4799]: E0930 14:46:22.751765 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fdc7b08_b5e8_45cb_97d9_d44658909edf.slice/crio-5167e632613ea6e9935e02cb1f76fbc9fd95ba43065afa177434c9005e783902\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fdc7b08_b5e8_45cb_97d9_d44658909edf.slice\": RecentStats: unable to find data in memory cache]" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.908029 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" event={"ID":"2fdc7b08-b5e8-45cb-97d9-d44658909edf","Type":"ContainerDied","Data":"5167e632613ea6e9935e02cb1f76fbc9fd95ba43065afa177434c9005e783902"} Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.908091 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5167e632613ea6e9935e02cb1f76fbc9fd95ba43065afa177434c9005e783902" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.908145 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7" Sep 30 14:46:22 crc kubenswrapper[4799]: I0930 14:46:22.999512 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55"] Sep 30 14:46:23 crc kubenswrapper[4799]: E0930 14:46:23.000114 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fdc7b08-b5e8-45cb-97d9-d44658909edf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.000146 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fdc7b08-b5e8-45cb-97d9-d44658909edf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.000397 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fdc7b08-b5e8-45cb-97d9-d44658909edf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.001258 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.003641 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.004027 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.004146 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.007155 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.012768 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55"] Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.145985 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.146307 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b96wx\" (UniqueName: \"kubernetes.io/projected/286e79d0-072f-422d-82cf-8932a5e9839f-kube-api-access-b96wx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.146415 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.248441 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.248756 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b96wx\" (UniqueName: \"kubernetes.io/projected/286e79d0-072f-422d-82cf-8932a5e9839f-kube-api-access-b96wx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.248796 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.253228 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.254256 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.267987 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b96wx\" (UniqueName: \"kubernetes.io/projected/286e79d0-072f-422d-82cf-8932a5e9839f-kube-api-access-b96wx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nwf55\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.323967 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.855899 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55"] Sep 30 14:46:23 crc kubenswrapper[4799]: I0930 14:46:23.921052 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" event={"ID":"286e79d0-072f-422d-82cf-8932a5e9839f","Type":"ContainerStarted","Data":"971b57a32b74a9c5db1585953082a9d84ac5251c3b290e2a45b10be9111c04fc"} Sep 30 14:46:24 crc kubenswrapper[4799]: I0930 14:46:24.503889 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:46:24 crc kubenswrapper[4799]: E0930 14:46:24.504528 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:46:24 crc kubenswrapper[4799]: I0930 14:46:24.935431 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" event={"ID":"286e79d0-072f-422d-82cf-8932a5e9839f","Type":"ContainerStarted","Data":"7f61815e7312a690b206b2310c1579444a98076f5112cab45dd329f3edde31b1"} Sep 30 14:46:24 crc kubenswrapper[4799]: I0930 14:46:24.959668 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" podStartSLOduration=2.78073103 podStartE2EDuration="2.95962755s" podCreationTimestamp="2025-09-30 14:46:22 +0000 UTC" firstStartedPulling="2025-09-30 14:46:23.862960919 +0000 UTC m=+1605.946561346" lastFinishedPulling="2025-09-30 14:46:24.041857439 +0000 UTC m=+1606.125457866" observedRunningTime="2025-09-30 14:46:24.958367623 +0000 UTC m=+1607.041968050" watchObservedRunningTime="2025-09-30 14:46:24.95962755 +0000 UTC m=+1607.043227977" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.421468 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5gg7d"] Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.425148 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.437664 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5gg7d"] Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.447000 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-catalog-content\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.466839 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-utilities\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.467213 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfj8s\" (UniqueName: \"kubernetes.io/projected/ce39374b-8cf6-472e-a0de-504c99fd7704-kube-api-access-nfj8s\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.571693 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-catalog-content\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.571773 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-utilities\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.571808 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfj8s\" (UniqueName: \"kubernetes.io/projected/ce39374b-8cf6-472e-a0de-504c99fd7704-kube-api-access-nfj8s\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.574319 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-utilities\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.574793 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-catalog-content\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.605845 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfj8s\" (UniqueName: \"kubernetes.io/projected/ce39374b-8cf6-472e-a0de-504c99fd7704-kube-api-access-nfj8s\") pod \"certified-operators-5gg7d\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.755931 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.976405 4799 generic.go:334] "Generic (PLEG): container finished" podID="286e79d0-072f-422d-82cf-8932a5e9839f" containerID="7f61815e7312a690b206b2310c1579444a98076f5112cab45dd329f3edde31b1" exitCode=0 Sep 30 14:46:27 crc kubenswrapper[4799]: I0930 14:46:27.976778 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" event={"ID":"286e79d0-072f-422d-82cf-8932a5e9839f","Type":"ContainerDied","Data":"7f61815e7312a690b206b2310c1579444a98076f5112cab45dd329f3edde31b1"} Sep 30 14:46:28 crc kubenswrapper[4799]: I0930 14:46:28.274209 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5gg7d"] Sep 30 14:46:28 crc kubenswrapper[4799]: W0930 14:46:28.277950 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce39374b_8cf6_472e_a0de_504c99fd7704.slice/crio-8074d3a15b7eaabb3ce78a693a5e91c8076a6dbee90dc11eec3db737d41cad66 WatchSource:0}: Error finding container 8074d3a15b7eaabb3ce78a693a5e91c8076a6dbee90dc11eec3db737d41cad66: Status 404 returned error can't find the container with id 8074d3a15b7eaabb3ce78a693a5e91c8076a6dbee90dc11eec3db737d41cad66 Sep 30 14:46:28 crc kubenswrapper[4799]: I0930 14:46:28.991416 4799 generic.go:334] "Generic (PLEG): container finished" podID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerID="01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2" exitCode=0 Sep 30 14:46:28 crc kubenswrapper[4799]: I0930 14:46:28.991486 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg7d" event={"ID":"ce39374b-8cf6-472e-a0de-504c99fd7704","Type":"ContainerDied","Data":"01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2"} Sep 30 14:46:28 crc kubenswrapper[4799]: I0930 14:46:28.991912 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg7d" event={"ID":"ce39374b-8cf6-472e-a0de-504c99fd7704","Type":"ContainerStarted","Data":"8074d3a15b7eaabb3ce78a693a5e91c8076a6dbee90dc11eec3db737d41cad66"} Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.456958 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.523429 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-ssh-key\") pod \"286e79d0-072f-422d-82cf-8932a5e9839f\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.523755 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-inventory\") pod \"286e79d0-072f-422d-82cf-8932a5e9839f\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.523879 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b96wx\" (UniqueName: \"kubernetes.io/projected/286e79d0-072f-422d-82cf-8932a5e9839f-kube-api-access-b96wx\") pod \"286e79d0-072f-422d-82cf-8932a5e9839f\" (UID: \"286e79d0-072f-422d-82cf-8932a5e9839f\") " Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.545284 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286e79d0-072f-422d-82cf-8932a5e9839f-kube-api-access-b96wx" (OuterVolumeSpecName: "kube-api-access-b96wx") pod "286e79d0-072f-422d-82cf-8932a5e9839f" (UID: "286e79d0-072f-422d-82cf-8932a5e9839f"). InnerVolumeSpecName "kube-api-access-b96wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.563512 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "286e79d0-072f-422d-82cf-8932a5e9839f" (UID: "286e79d0-072f-422d-82cf-8932a5e9839f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.573984 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-inventory" (OuterVolumeSpecName: "inventory") pod "286e79d0-072f-422d-82cf-8932a5e9839f" (UID: "286e79d0-072f-422d-82cf-8932a5e9839f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.627587 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b96wx\" (UniqueName: \"kubernetes.io/projected/286e79d0-072f-422d-82cf-8932a5e9839f-kube-api-access-b96wx\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.627836 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:29 crc kubenswrapper[4799]: I0930 14:46:29.628000 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/286e79d0-072f-422d-82cf-8932a5e9839f-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.005394 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg7d" event={"ID":"ce39374b-8cf6-472e-a0de-504c99fd7704","Type":"ContainerStarted","Data":"7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1"} Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.007959 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" event={"ID":"286e79d0-072f-422d-82cf-8932a5e9839f","Type":"ContainerDied","Data":"971b57a32b74a9c5db1585953082a9d84ac5251c3b290e2a45b10be9111c04fc"} Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.007989 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="971b57a32b74a9c5db1585953082a9d84ac5251c3b290e2a45b10be9111c04fc" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.008062 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nwf55" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.112004 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl"] Sep 30 14:46:30 crc kubenswrapper[4799]: E0930 14:46:30.113501 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286e79d0-072f-422d-82cf-8932a5e9839f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.113531 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="286e79d0-072f-422d-82cf-8932a5e9839f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.142731 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="286e79d0-072f-422d-82cf-8932a5e9839f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.144224 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.147547 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl"] Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.148039 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.150566 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.152185 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.152241 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.283124 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcr26\" (UniqueName: \"kubernetes.io/projected/4565a6db-10d1-440b-a7b7-dc338c18f70f-kube-api-access-zcr26\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.283504 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.283540 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.283731 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.385431 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.385535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcr26\" (UniqueName: \"kubernetes.io/projected/4565a6db-10d1-440b-a7b7-dc338c18f70f-kube-api-access-zcr26\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.385615 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.385644 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.391082 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.391757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.394032 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.408584 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcr26\" (UniqueName: \"kubernetes.io/projected/4565a6db-10d1-440b-a7b7-dc338c18f70f-kube-api-access-zcr26\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:30 crc kubenswrapper[4799]: I0930 14:46:30.521552 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:46:31 crc kubenswrapper[4799]: I0930 14:46:31.166945 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl"] Sep 30 14:46:32 crc kubenswrapper[4799]: I0930 14:46:32.036392 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" event={"ID":"4565a6db-10d1-440b-a7b7-dc338c18f70f","Type":"ContainerStarted","Data":"757f4446877ae3c1292ee94ce214a3bd548edbe4f43f3f3b82ed4da01c8f6a36"} Sep 30 14:46:32 crc kubenswrapper[4799]: I0930 14:46:32.036855 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" event={"ID":"4565a6db-10d1-440b-a7b7-dc338c18f70f","Type":"ContainerStarted","Data":"41ae7a858fef07f1fec190e02dd09c39133158b7f1457bc1ffb24c9c0e577c1e"} Sep 30 14:46:32 crc kubenswrapper[4799]: I0930 14:46:32.058821 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" podStartSLOduration=1.8284711919999999 podStartE2EDuration="2.058753984s" podCreationTimestamp="2025-09-30 14:46:30 +0000 UTC" firstStartedPulling="2025-09-30 14:46:31.17846581 +0000 UTC m=+1613.262066237" lastFinishedPulling="2025-09-30 14:46:31.408748602 +0000 UTC m=+1613.492349029" observedRunningTime="2025-09-30 14:46:32.057062855 +0000 UTC m=+1614.140663502" watchObservedRunningTime="2025-09-30 14:46:32.058753984 +0000 UTC m=+1614.142354421" Sep 30 14:46:33 crc kubenswrapper[4799]: I0930 14:46:33.051702 4799 generic.go:334] "Generic (PLEG): container finished" podID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerID="7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1" exitCode=0 Sep 30 14:46:33 crc kubenswrapper[4799]: I0930 14:46:33.051793 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg7d" event={"ID":"ce39374b-8cf6-472e-a0de-504c99fd7704","Type":"ContainerDied","Data":"7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1"} Sep 30 14:46:34 crc kubenswrapper[4799]: I0930 14:46:34.064215 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg7d" event={"ID":"ce39374b-8cf6-472e-a0de-504c99fd7704","Type":"ContainerStarted","Data":"788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9"} Sep 30 14:46:34 crc kubenswrapper[4799]: I0930 14:46:34.086895 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5gg7d" podStartSLOduration=2.460189717 podStartE2EDuration="7.086869013s" podCreationTimestamp="2025-09-30 14:46:27 +0000 UTC" firstStartedPulling="2025-09-30 14:46:28.994945836 +0000 UTC m=+1611.078546263" lastFinishedPulling="2025-09-30 14:46:33.621625132 +0000 UTC m=+1615.705225559" observedRunningTime="2025-09-30 14:46:34.084193825 +0000 UTC m=+1616.167794272" watchObservedRunningTime="2025-09-30 14:46:34.086869013 +0000 UTC m=+1616.170469440" Sep 30 14:46:37 crc kubenswrapper[4799]: I0930 14:46:37.505043 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:46:37 crc kubenswrapper[4799]: E0930 14:46:37.505626 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:46:37 crc kubenswrapper[4799]: I0930 14:46:37.756765 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:37 crc kubenswrapper[4799]: I0930 14:46:37.757251 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:38 crc kubenswrapper[4799]: I0930 14:46:38.810755 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-5gg7d" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="registry-server" probeResult="failure" output=< Sep 30 14:46:38 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 14:46:38 crc kubenswrapper[4799]: > Sep 30 14:46:47 crc kubenswrapper[4799]: I0930 14:46:47.811512 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:47 crc kubenswrapper[4799]: I0930 14:46:47.869963 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:48 crc kubenswrapper[4799]: I0930 14:46:48.058248 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5gg7d"] Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.219181 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5gg7d" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="registry-server" containerID="cri-o://788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9" gracePeriod=2 Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.506793 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:46:49 crc kubenswrapper[4799]: E0930 14:46:49.507406 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.718248 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.828712 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfj8s\" (UniqueName: \"kubernetes.io/projected/ce39374b-8cf6-472e-a0de-504c99fd7704-kube-api-access-nfj8s\") pod \"ce39374b-8cf6-472e-a0de-504c99fd7704\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.829370 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-catalog-content\") pod \"ce39374b-8cf6-472e-a0de-504c99fd7704\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.829529 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-utilities\") pod \"ce39374b-8cf6-472e-a0de-504c99fd7704\" (UID: \"ce39374b-8cf6-472e-a0de-504c99fd7704\") " Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.830531 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-utilities" (OuterVolumeSpecName: "utilities") pod "ce39374b-8cf6-472e-a0de-504c99fd7704" (UID: "ce39374b-8cf6-472e-a0de-504c99fd7704"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.836514 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce39374b-8cf6-472e-a0de-504c99fd7704-kube-api-access-nfj8s" (OuterVolumeSpecName: "kube-api-access-nfj8s") pod "ce39374b-8cf6-472e-a0de-504c99fd7704" (UID: "ce39374b-8cf6-472e-a0de-504c99fd7704"). InnerVolumeSpecName "kube-api-access-nfj8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.879642 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce39374b-8cf6-472e-a0de-504c99fd7704" (UID: "ce39374b-8cf6-472e-a0de-504c99fd7704"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.931788 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.931822 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce39374b-8cf6-472e-a0de-504c99fd7704-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:49 crc kubenswrapper[4799]: I0930 14:46:49.931833 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfj8s\" (UniqueName: \"kubernetes.io/projected/ce39374b-8cf6-472e-a0de-504c99fd7704-kube-api-access-nfj8s\") on node \"crc\" DevicePath \"\"" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.234446 4799 generic.go:334] "Generic (PLEG): container finished" podID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerID="788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9" exitCode=0 Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.234512 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg7d" event={"ID":"ce39374b-8cf6-472e-a0de-504c99fd7704","Type":"ContainerDied","Data":"788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9"} Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.234551 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg7d" event={"ID":"ce39374b-8cf6-472e-a0de-504c99fd7704","Type":"ContainerDied","Data":"8074d3a15b7eaabb3ce78a693a5e91c8076a6dbee90dc11eec3db737d41cad66"} Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.234576 4799 scope.go:117] "RemoveContainer" containerID="788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.234797 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gg7d" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.292587 4799 scope.go:117] "RemoveContainer" containerID="7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.298020 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5gg7d"] Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.312770 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5gg7d"] Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.335389 4799 scope.go:117] "RemoveContainer" containerID="01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.395672 4799 scope.go:117] "RemoveContainer" containerID="788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9" Sep 30 14:46:50 crc kubenswrapper[4799]: E0930 14:46:50.396397 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9\": container with ID starting with 788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9 not found: ID does not exist" containerID="788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.396591 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9"} err="failed to get container status \"788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9\": rpc error: code = NotFound desc = could not find container \"788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9\": container with ID starting with 788607c2ceba7324fe411da9cd5990a40b9a4525e45c1115d812c06e504895b9 not found: ID does not exist" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.396877 4799 scope.go:117] "RemoveContainer" containerID="7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1" Sep 30 14:46:50 crc kubenswrapper[4799]: E0930 14:46:50.397575 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1\": container with ID starting with 7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1 not found: ID does not exist" containerID="7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.397621 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1"} err="failed to get container status \"7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1\": rpc error: code = NotFound desc = could not find container \"7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1\": container with ID starting with 7e355589fbfa0ade13e8ee0289a42acb68c2d7bd0593649a5366f24a546e0cc1 not found: ID does not exist" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.397672 4799 scope.go:117] "RemoveContainer" containerID="01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2" Sep 30 14:46:50 crc kubenswrapper[4799]: E0930 14:46:50.398019 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2\": container with ID starting with 01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2 not found: ID does not exist" containerID="01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.398167 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2"} err="failed to get container status \"01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2\": rpc error: code = NotFound desc = could not find container \"01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2\": container with ID starting with 01856b7d7c2ec4d886dbc552bff041fd14872d7fe2bbc3c10333fa0aa85fadc2 not found: ID does not exist" Sep 30 14:46:50 crc kubenswrapper[4799]: I0930 14:46:50.515932 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" path="/var/lib/kubelet/pods/ce39374b-8cf6-472e-a0de-504c99fd7704/volumes" Sep 30 14:47:03 crc kubenswrapper[4799]: I0930 14:47:03.504012 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:47:03 crc kubenswrapper[4799]: E0930 14:47:03.505264 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:47:08 crc kubenswrapper[4799]: I0930 14:47:08.923752 4799 scope.go:117] "RemoveContainer" containerID="0dbe8d62cbbbfc49cb340f4e23a3440773b4a36fb9035d93f4596e3bb41f36fd" Sep 30 14:47:08 crc kubenswrapper[4799]: I0930 14:47:08.949802 4799 scope.go:117] "RemoveContainer" containerID="8923b064d4910eddccd36258c0091f291d87f5d2b5ba8a2a2855d3bafc5f9665" Sep 30 14:47:08 crc kubenswrapper[4799]: I0930 14:47:08.986291 4799 scope.go:117] "RemoveContainer" containerID="5cb5e6829b3afc23b8e4bc1f2df067901aa9c895fc1897c6133fb7845aa14fe7" Sep 30 14:47:16 crc kubenswrapper[4799]: I0930 14:47:16.504606 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:47:16 crc kubenswrapper[4799]: E0930 14:47:16.506669 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:47:30 crc kubenswrapper[4799]: I0930 14:47:30.503687 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:47:30 crc kubenswrapper[4799]: E0930 14:47:30.505837 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:47:42 crc kubenswrapper[4799]: I0930 14:47:42.506053 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:47:42 crc kubenswrapper[4799]: E0930 14:47:42.506882 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:47:55 crc kubenswrapper[4799]: I0930 14:47:55.505878 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:47:55 crc kubenswrapper[4799]: E0930 14:47:55.508011 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:48:09 crc kubenswrapper[4799]: I0930 14:48:09.104969 4799 scope.go:117] "RemoveContainer" containerID="e400d2c05e69e3589bef5450ad1addec29192ec4857b5fd384600d77599bbe3d" Sep 30 14:48:09 crc kubenswrapper[4799]: I0930 14:48:09.132737 4799 scope.go:117] "RemoveContainer" containerID="32fa7121b9a6200b71434ee14f58f183fde97f26fa021511cdaf7794213b491d" Sep 30 14:48:09 crc kubenswrapper[4799]: I0930 14:48:09.157939 4799 scope.go:117] "RemoveContainer" containerID="713617e71c53f603a2a075b33b4a2d4f713d8b1d2377b5c739ca098e76d3c79f" Sep 30 14:48:09 crc kubenswrapper[4799]: I0930 14:48:09.183398 4799 scope.go:117] "RemoveContainer" containerID="fa48171e870ac0e6c827755093aa34ec1156ebefb32dad88d1cafdc5f78e10a0" Sep 30 14:48:10 crc kubenswrapper[4799]: I0930 14:48:10.504366 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:48:10 crc kubenswrapper[4799]: E0930 14:48:10.504672 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:48:13 crc kubenswrapper[4799]: I0930 14:48:13.045168 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-4tjtd"] Sep 30 14:48:13 crc kubenswrapper[4799]: I0930 14:48:13.051951 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-4tjtd"] Sep 30 14:48:14 crc kubenswrapper[4799]: I0930 14:48:14.041230 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-rqtlt"] Sep 30 14:48:14 crc kubenswrapper[4799]: I0930 14:48:14.049962 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-rqtlt"] Sep 30 14:48:14 crc kubenswrapper[4799]: I0930 14:48:14.557523 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6f59e8e-8173-468f-81c8-9f86067de6ea" path="/var/lib/kubelet/pods/c6f59e8e-8173-468f-81c8-9f86067de6ea/volumes" Sep 30 14:48:14 crc kubenswrapper[4799]: I0930 14:48:14.560571 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccab43ff-518e-4785-9234-477293d0da8b" path="/var/lib/kubelet/pods/ccab43ff-518e-4785-9234-477293d0da8b/volumes" Sep 30 14:48:18 crc kubenswrapper[4799]: I0930 14:48:18.030509 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-kv6wf"] Sep 30 14:48:18 crc kubenswrapper[4799]: I0930 14:48:18.041167 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-kv6wf"] Sep 30 14:48:18 crc kubenswrapper[4799]: I0930 14:48:18.530766 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b" path="/var/lib/kubelet/pods/4ec24c32-488b-41b9-b8f9-f4cc4cb67c0b/volumes" Sep 30 14:48:22 crc kubenswrapper[4799]: I0930 14:48:22.040324 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b430-account-create-bh8t8"] Sep 30 14:48:22 crc kubenswrapper[4799]: I0930 14:48:22.051387 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b430-account-create-bh8t8"] Sep 30 14:48:22 crc kubenswrapper[4799]: I0930 14:48:22.515530 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d63ec7b1-a806-40c4-9f60-a17b892e27cf" path="/var/lib/kubelet/pods/d63ec7b1-a806-40c4-9f60-a17b892e27cf/volumes" Sep 30 14:48:23 crc kubenswrapper[4799]: I0930 14:48:23.503405 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:48:23 crc kubenswrapper[4799]: E0930 14:48:23.503743 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:48:27 crc kubenswrapper[4799]: I0930 14:48:27.063744 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-9117-account-create-qfjcx"] Sep 30 14:48:27 crc kubenswrapper[4799]: I0930 14:48:27.073484 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-9117-account-create-qfjcx"] Sep 30 14:48:28 crc kubenswrapper[4799]: I0930 14:48:28.519543 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0d8d726-9fbf-4a54-bcd1-835dc062709f" path="/var/lib/kubelet/pods/a0d8d726-9fbf-4a54-bcd1-835dc062709f/volumes" Sep 30 14:48:34 crc kubenswrapper[4799]: I0930 14:48:34.034874 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-8a5a-account-create-kmzh9"] Sep 30 14:48:34 crc kubenswrapper[4799]: I0930 14:48:34.043545 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-8a5a-account-create-kmzh9"] Sep 30 14:48:34 crc kubenswrapper[4799]: I0930 14:48:34.523098 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de063760-02b1-4dae-b475-8e33b7f5bf66" path="/var/lib/kubelet/pods/de063760-02b1-4dae-b475-8e33b7f5bf66/volumes" Sep 30 14:48:36 crc kubenswrapper[4799]: I0930 14:48:36.504199 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:48:36 crc kubenswrapper[4799]: E0930 14:48:36.506080 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:48:48 crc kubenswrapper[4799]: I0930 14:48:48.510427 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:48:48 crc kubenswrapper[4799]: E0930 14:48:48.511552 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:48:53 crc kubenswrapper[4799]: I0930 14:48:53.032232 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9gk2n"] Sep 30 14:48:53 crc kubenswrapper[4799]: I0930 14:48:53.042242 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-m88q9"] Sep 30 14:48:53 crc kubenswrapper[4799]: I0930 14:48:53.057298 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-7gkqb"] Sep 30 14:48:53 crc kubenswrapper[4799]: I0930 14:48:53.066000 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9gk2n"] Sep 30 14:48:53 crc kubenswrapper[4799]: I0930 14:48:53.075194 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-m88q9"] Sep 30 14:48:53 crc kubenswrapper[4799]: I0930 14:48:53.083607 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-7gkqb"] Sep 30 14:48:54 crc kubenswrapper[4799]: I0930 14:48:54.517213 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5da50628-e5d7-42e6-a221-f8503afc6885" path="/var/lib/kubelet/pods/5da50628-e5d7-42e6-a221-f8503afc6885/volumes" Sep 30 14:48:54 crc kubenswrapper[4799]: I0930 14:48:54.519557 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e35e950-34b7-4736-b669-6d3c52d0c797" path="/var/lib/kubelet/pods/5e35e950-34b7-4736-b669-6d3c52d0c797/volumes" Sep 30 14:48:54 crc kubenswrapper[4799]: I0930 14:48:54.521807 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bb77af6-6c03-44f8-bec8-d168840a0c1e" path="/var/lib/kubelet/pods/8bb77af6-6c03-44f8-bec8-d168840a0c1e/volumes" Sep 30 14:49:02 crc kubenswrapper[4799]: I0930 14:49:02.503671 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:49:02 crc kubenswrapper[4799]: E0930 14:49:02.504527 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:49:08 crc kubenswrapper[4799]: I0930 14:49:08.044998 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4d9n4"] Sep 30 14:49:08 crc kubenswrapper[4799]: I0930 14:49:08.052614 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4d9n4"] Sep 30 14:49:08 crc kubenswrapper[4799]: I0930 14:49:08.515041 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfbbeb1f-d64a-46ac-a80d-3541ee813f6b" path="/var/lib/kubelet/pods/bfbbeb1f-d64a-46ac-a80d-3541ee813f6b/volumes" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.243243 4799 scope.go:117] "RemoveContainer" containerID="3ba1f87126afcc820da7f7ca3ba6d11ef72a1bc8d3a4d7c1f99d838c77ee19b2" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.273918 4799 scope.go:117] "RemoveContainer" containerID="d4f5e3d31e35f437fd7a5ea9259d85f3d7be6c3703cdd068696ab1a4084a0e5e" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.347289 4799 scope.go:117] "RemoveContainer" containerID="a70bfcac89f1ba8547ee90d38bc9a91f00495ceea0a895028b8bec5c8e914cfe" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.374843 4799 scope.go:117] "RemoveContainer" containerID="b9c74512e48d404333cddb32ef0922c45770ad28d46a1af0a7d44a46a4307d38" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.424185 4799 scope.go:117] "RemoveContainer" containerID="987166546ed56ca282df48dcdde95734d12003ca61d52d68a574dd3ee10b92fb" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.482564 4799 scope.go:117] "RemoveContainer" containerID="f2cebab7abd7a7cc43f8ebbf8af08ab2a140cb6467343873cc3c86e255405695" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.533712 4799 scope.go:117] "RemoveContainer" containerID="5979ea5bf03171e1964c75cd98fc7c15dfb205529d113bc141f6a5b170af3f6e" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.588615 4799 scope.go:117] "RemoveContainer" containerID="ee11a34e8a354858a54f65100e7546d1995e2abb7f1f4561c6a64358e4595b1c" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.620878 4799 scope.go:117] "RemoveContainer" containerID="cee3108a6408931221e66323254f6d4dc416ae87a8cf6f46f06f34655a58df6f" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.648161 4799 scope.go:117] "RemoveContainer" containerID="9180d755716119ac42d286c0e8752d52e15d1321c059368d6fd7c955d3700d5c" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.674212 4799 scope.go:117] "RemoveContainer" containerID="e122a9b6ce81c231cea3fbcecb5741d02ff6e0b0378223608e5cda96a1b17d76" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.702885 4799 scope.go:117] "RemoveContainer" containerID="a17141d687adedf845311a4a394b436b1c880cf58ca3b23dc94d4706d2346cbf" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.736506 4799 scope.go:117] "RemoveContainer" containerID="b0444f5ffa332872dc4acbe21d5b3e146da1d0ae4f8449809fccef5b55d119f9" Sep 30 14:49:09 crc kubenswrapper[4799]: I0930 14:49:09.760917 4799 scope.go:117] "RemoveContainer" containerID="34e9f5d1a8a89605dc771a739f141bfb3191b1f19adba7d51b71aa26eae4c0ab" Sep 30 14:49:12 crc kubenswrapper[4799]: I0930 14:49:12.033484 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ac6f-account-create-cfpcr"] Sep 30 14:49:12 crc kubenswrapper[4799]: I0930 14:49:12.043736 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ac6f-account-create-cfpcr"] Sep 30 14:49:12 crc kubenswrapper[4799]: I0930 14:49:12.058402 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6f25-account-create-hcvsn"] Sep 30 14:49:12 crc kubenswrapper[4799]: I0930 14:49:12.069208 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6f25-account-create-hcvsn"] Sep 30 14:49:12 crc kubenswrapper[4799]: I0930 14:49:12.516162 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acec8647-568b-474c-86a6-590726b7b86e" path="/var/lib/kubelet/pods/acec8647-568b-474c-86a6-590726b7b86e/volumes" Sep 30 14:49:12 crc kubenswrapper[4799]: I0930 14:49:12.516890 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea31ca1c-9064-41a9-a55e-24d09ee891f3" path="/var/lib/kubelet/pods/ea31ca1c-9064-41a9-a55e-24d09ee891f3/volumes" Sep 30 14:49:13 crc kubenswrapper[4799]: I0930 14:49:13.034089 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-4c10-account-create-r5fpd"] Sep 30 14:49:13 crc kubenswrapper[4799]: I0930 14:49:13.048568 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-4c10-account-create-r5fpd"] Sep 30 14:49:14 crc kubenswrapper[4799]: I0930 14:49:14.033040 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-nxr4c"] Sep 30 14:49:14 crc kubenswrapper[4799]: I0930 14:49:14.038868 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-nxr4c"] Sep 30 14:49:14 crc kubenswrapper[4799]: I0930 14:49:14.520892 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499" path="/var/lib/kubelet/pods/a0bee0b4-7c57-4529-b6dc-ca8e9c4ba499/volumes" Sep 30 14:49:14 crc kubenswrapper[4799]: I0930 14:49:14.523624 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3469c1a-9cfa-4d09-9672-40de98632e0e" path="/var/lib/kubelet/pods/d3469c1a-9cfa-4d09-9672-40de98632e0e/volumes" Sep 30 14:49:17 crc kubenswrapper[4799]: I0930 14:49:17.503038 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:49:17 crc kubenswrapper[4799]: E0930 14:49:17.503711 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:49:32 crc kubenswrapper[4799]: I0930 14:49:32.503772 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:49:32 crc kubenswrapper[4799]: E0930 14:49:32.504594 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:49:45 crc kubenswrapper[4799]: I0930 14:49:45.503360 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:49:45 crc kubenswrapper[4799]: E0930 14:49:45.504439 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:49:51 crc kubenswrapper[4799]: I0930 14:49:51.170684 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" event={"ID":"4565a6db-10d1-440b-a7b7-dc338c18f70f","Type":"ContainerDied","Data":"757f4446877ae3c1292ee94ce214a3bd548edbe4f43f3f3b82ed4da01c8f6a36"} Sep 30 14:49:51 crc kubenswrapper[4799]: I0930 14:49:51.170677 4799 generic.go:334] "Generic (PLEG): container finished" podID="4565a6db-10d1-440b-a7b7-dc338c18f70f" containerID="757f4446877ae3c1292ee94ce214a3bd548edbe4f43f3f3b82ed4da01c8f6a36" exitCode=0 Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.640113 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.795626 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-bootstrap-combined-ca-bundle\") pod \"4565a6db-10d1-440b-a7b7-dc338c18f70f\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.795701 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcr26\" (UniqueName: \"kubernetes.io/projected/4565a6db-10d1-440b-a7b7-dc338c18f70f-kube-api-access-zcr26\") pod \"4565a6db-10d1-440b-a7b7-dc338c18f70f\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.795764 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-ssh-key\") pod \"4565a6db-10d1-440b-a7b7-dc338c18f70f\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.795823 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-inventory\") pod \"4565a6db-10d1-440b-a7b7-dc338c18f70f\" (UID: \"4565a6db-10d1-440b-a7b7-dc338c18f70f\") " Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.834972 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4565a6db-10d1-440b-a7b7-dc338c18f70f-kube-api-access-zcr26" (OuterVolumeSpecName: "kube-api-access-zcr26") pod "4565a6db-10d1-440b-a7b7-dc338c18f70f" (UID: "4565a6db-10d1-440b-a7b7-dc338c18f70f"). InnerVolumeSpecName "kube-api-access-zcr26". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.839903 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4565a6db-10d1-440b-a7b7-dc338c18f70f" (UID: "4565a6db-10d1-440b-a7b7-dc338c18f70f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.842813 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4565a6db-10d1-440b-a7b7-dc338c18f70f" (UID: "4565a6db-10d1-440b-a7b7-dc338c18f70f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.848884 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-inventory" (OuterVolumeSpecName: "inventory") pod "4565a6db-10d1-440b-a7b7-dc338c18f70f" (UID: "4565a6db-10d1-440b-a7b7-dc338c18f70f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.898116 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.898162 4799 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.898179 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcr26\" (UniqueName: \"kubernetes.io/projected/4565a6db-10d1-440b-a7b7-dc338c18f70f-kube-api-access-zcr26\") on node \"crc\" DevicePath \"\"" Sep 30 14:49:52 crc kubenswrapper[4799]: I0930 14:49:52.898187 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4565a6db-10d1-440b-a7b7-dc338c18f70f-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.190735 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" event={"ID":"4565a6db-10d1-440b-a7b7-dc338c18f70f","Type":"ContainerDied","Data":"41ae7a858fef07f1fec190e02dd09c39133158b7f1457bc1ffb24c9c0e577c1e"} Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.190786 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41ae7a858fef07f1fec190e02dd09c39133158b7f1457bc1ffb24c9c0e577c1e" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.190829 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.308995 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv"] Sep 30 14:49:53 crc kubenswrapper[4799]: E0930 14:49:53.309427 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="extract-utilities" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.309444 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="extract-utilities" Sep 30 14:49:53 crc kubenswrapper[4799]: E0930 14:49:53.309452 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="extract-content" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.309458 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="extract-content" Sep 30 14:49:53 crc kubenswrapper[4799]: E0930 14:49:53.309494 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4565a6db-10d1-440b-a7b7-dc338c18f70f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.309501 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4565a6db-10d1-440b-a7b7-dc338c18f70f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 14:49:53 crc kubenswrapper[4799]: E0930 14:49:53.309512 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="registry-server" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.309518 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="registry-server" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.309733 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4565a6db-10d1-440b-a7b7-dc338c18f70f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.309750 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce39374b-8cf6-472e-a0de-504c99fd7704" containerName="registry-server" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.310463 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.313284 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.313328 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.313473 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.313566 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.331173 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv"] Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.407399 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl42h\" (UniqueName: \"kubernetes.io/projected/8da29a07-04e5-4c03-adf7-f642c86abecd-kube-api-access-zl42h\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.407469 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.407497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.511110 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl42h\" (UniqueName: \"kubernetes.io/projected/8da29a07-04e5-4c03-adf7-f642c86abecd-kube-api-access-zl42h\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.511300 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.511338 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.515376 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.515411 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.543792 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl42h\" (UniqueName: \"kubernetes.io/projected/8da29a07-04e5-4c03-adf7-f642c86abecd-kube-api-access-zl42h\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-svwrv\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:53 crc kubenswrapper[4799]: I0930 14:49:53.631163 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:49:54 crc kubenswrapper[4799]: I0930 14:49:54.185028 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv"] Sep 30 14:49:54 crc kubenswrapper[4799]: W0930 14:49:54.198140 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8da29a07_04e5_4c03_adf7_f642c86abecd.slice/crio-9972227e39af6dd0c015641195ee8f26548c411d7c9c1784b5a0105377875c41 WatchSource:0}: Error finding container 9972227e39af6dd0c015641195ee8f26548c411d7c9c1784b5a0105377875c41: Status 404 returned error can't find the container with id 9972227e39af6dd0c015641195ee8f26548c411d7c9c1784b5a0105377875c41 Sep 30 14:49:54 crc kubenswrapper[4799]: I0930 14:49:54.204517 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:49:54 crc kubenswrapper[4799]: I0930 14:49:54.221717 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" event={"ID":"8da29a07-04e5-4c03-adf7-f642c86abecd","Type":"ContainerStarted","Data":"9972227e39af6dd0c015641195ee8f26548c411d7c9c1784b5a0105377875c41"} Sep 30 14:49:55 crc kubenswrapper[4799]: I0930 14:49:55.235822 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" event={"ID":"8da29a07-04e5-4c03-adf7-f642c86abecd","Type":"ContainerStarted","Data":"4d1c94b5447917d2b8d411a53cab0f977fe5acb19a7c24f759e1c2f732b67bd6"} Sep 30 14:49:55 crc kubenswrapper[4799]: I0930 14:49:55.263309 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" podStartSLOduration=2.065137962 podStartE2EDuration="2.263280943s" podCreationTimestamp="2025-09-30 14:49:53 +0000 UTC" firstStartedPulling="2025-09-30 14:49:54.204195262 +0000 UTC m=+1816.287795689" lastFinishedPulling="2025-09-30 14:49:54.402338253 +0000 UTC m=+1816.485938670" observedRunningTime="2025-09-30 14:49:55.257918597 +0000 UTC m=+1817.341519034" watchObservedRunningTime="2025-09-30 14:49:55.263280943 +0000 UTC m=+1817.346881370" Sep 30 14:50:00 crc kubenswrapper[4799]: I0930 14:50:00.505202 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:50:00 crc kubenswrapper[4799]: E0930 14:50:00.505988 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:50:03 crc kubenswrapper[4799]: I0930 14:50:03.050962 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-fb472"] Sep 30 14:50:03 crc kubenswrapper[4799]: I0930 14:50:03.060880 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-fb472"] Sep 30 14:50:04 crc kubenswrapper[4799]: I0930 14:50:04.519420 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e3be2e-5fd3-4ac7-9138-c968d3d0c025" path="/var/lib/kubelet/pods/67e3be2e-5fd3-4ac7-9138-c968d3d0c025/volumes" Sep 30 14:50:10 crc kubenswrapper[4799]: I0930 14:50:10.039888 4799 scope.go:117] "RemoveContainer" containerID="d3f69fa4bdfb6650ca83eb0148fd6757e7c8a958c66b2c3af14ed14c089a5189" Sep 30 14:50:10 crc kubenswrapper[4799]: I0930 14:50:10.068899 4799 scope.go:117] "RemoveContainer" containerID="85994786d1236ad433f7cb78913926fed48b7e24160914eb8d1d855e9768c479" Sep 30 14:50:10 crc kubenswrapper[4799]: I0930 14:50:10.131919 4799 scope.go:117] "RemoveContainer" containerID="a7cb6d772c2f7b01603f66796135de6d05559ae3bef5e92adda80722fa0bd601" Sep 30 14:50:10 crc kubenswrapper[4799]: I0930 14:50:10.179998 4799 scope.go:117] "RemoveContainer" containerID="413c359b00855aeb89ae340b26ae9452dce56269bd92cb960e5dffc6735e24d2" Sep 30 14:50:10 crc kubenswrapper[4799]: I0930 14:50:10.225953 4799 scope.go:117] "RemoveContainer" containerID="a0efc0d12665083a1638a4b9241694233639c17b85a2484725aa6a18f922f2b7" Sep 30 14:50:12 crc kubenswrapper[4799]: I0930 14:50:12.038165 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qrs4k"] Sep 30 14:50:12 crc kubenswrapper[4799]: I0930 14:50:12.048584 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qrs4k"] Sep 30 14:50:12 crc kubenswrapper[4799]: I0930 14:50:12.516248 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e71e7667-1d26-47fc-9884-1c946e72c244" path="/var/lib/kubelet/pods/e71e7667-1d26-47fc-9884-1c946e72c244/volumes" Sep 30 14:50:15 crc kubenswrapper[4799]: I0930 14:50:15.503427 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:50:15 crc kubenswrapper[4799]: E0930 14:50:15.503956 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:50:16 crc kubenswrapper[4799]: I0930 14:50:16.043846 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-l7xw2"] Sep 30 14:50:16 crc kubenswrapper[4799]: I0930 14:50:16.053133 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-l7xw2"] Sep 30 14:50:16 crc kubenswrapper[4799]: I0930 14:50:16.516580 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229" path="/var/lib/kubelet/pods/0c9ce36c-5f9d-43b2-8b0c-1cf52eb40229/volumes" Sep 30 14:50:27 crc kubenswrapper[4799]: I0930 14:50:27.503581 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:50:27 crc kubenswrapper[4799]: E0930 14:50:27.504593 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:50:29 crc kubenswrapper[4799]: I0930 14:50:29.091078 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xt6j5"] Sep 30 14:50:29 crc kubenswrapper[4799]: I0930 14:50:29.101181 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xt6j5"] Sep 30 14:50:30 crc kubenswrapper[4799]: I0930 14:50:30.517948 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d30b6ec-a2d1-4f93-932b-e78856253cc9" path="/var/lib/kubelet/pods/2d30b6ec-a2d1-4f93-932b-e78856253cc9/volumes" Sep 30 14:50:34 crc kubenswrapper[4799]: I0930 14:50:34.039450 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-mx9fz"] Sep 30 14:50:34 crc kubenswrapper[4799]: I0930 14:50:34.052772 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-mx9fz"] Sep 30 14:50:34 crc kubenswrapper[4799]: I0930 14:50:34.518844 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78ac655a-3fc7-46a4-93c2-69574e8b8f22" path="/var/lib/kubelet/pods/78ac655a-3fc7-46a4-93c2-69574e8b8f22/volumes" Sep 30 14:50:42 crc kubenswrapper[4799]: I0930 14:50:42.504916 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:50:43 crc kubenswrapper[4799]: I0930 14:50:43.777973 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"116aa297eadbb7887d7d899e95966cb2028cb7fa70dc029dd1713cd3c032c2a4"} Sep 30 14:51:10 crc kubenswrapper[4799]: I0930 14:51:10.415051 4799 scope.go:117] "RemoveContainer" containerID="87839f98445eabda60d578fba0d8f5cbc36fdf4a86b85aaf163eaa53acb92000" Sep 30 14:51:10 crc kubenswrapper[4799]: I0930 14:51:10.453534 4799 scope.go:117] "RemoveContainer" containerID="6f7d06a08493951cd016e08396484bd16fb6f4e1463d60160b07330decf52bfb" Sep 30 14:51:10 crc kubenswrapper[4799]: I0930 14:51:10.498448 4799 scope.go:117] "RemoveContainer" containerID="15813eeea4cf4efe8492111998588d8903edae8edcbbf0a1ab40735200a85876" Sep 30 14:51:10 crc kubenswrapper[4799]: I0930 14:51:10.713969 4799 scope.go:117] "RemoveContainer" containerID="a86ae59bf3e2020b35f1021225e2b995c371d1ee708c374d3038fd248b6dcc86" Sep 30 14:51:10 crc kubenswrapper[4799]: I0930 14:51:10.752013 4799 scope.go:117] "RemoveContainer" containerID="7f62f4f5059123bfcf58b12347dd2f0cb1506418ef5092f072080f309570c950" Sep 30 14:51:33 crc kubenswrapper[4799]: I0930 14:51:33.047830 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-s9qbm"] Sep 30 14:51:33 crc kubenswrapper[4799]: I0930 14:51:33.055953 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-fpsgw"] Sep 30 14:51:33 crc kubenswrapper[4799]: I0930 14:51:33.063533 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-74vk7"] Sep 30 14:51:33 crc kubenswrapper[4799]: I0930 14:51:33.071425 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-74vk7"] Sep 30 14:51:33 crc kubenswrapper[4799]: I0930 14:51:33.080451 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-fpsgw"] Sep 30 14:51:33 crc kubenswrapper[4799]: I0930 14:51:33.091277 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-s9qbm"] Sep 30 14:51:34 crc kubenswrapper[4799]: I0930 14:51:34.517134 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99b988bd-7bdb-455e-a946-0586b3be1e7c" path="/var/lib/kubelet/pods/99b988bd-7bdb-455e-a946-0586b3be1e7c/volumes" Sep 30 14:51:34 crc kubenswrapper[4799]: I0930 14:51:34.518584 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17315e1-62f6-4402-9b3e-90a88c4beaba" path="/var/lib/kubelet/pods/d17315e1-62f6-4402-9b3e-90a88c4beaba/volumes" Sep 30 14:51:34 crc kubenswrapper[4799]: I0930 14:51:34.519281 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcfcac79-a568-4c8a-aff4-5533d18d954d" path="/var/lib/kubelet/pods/fcfcac79-a568-4c8a-aff4-5533d18d954d/volumes" Sep 30 14:51:39 crc kubenswrapper[4799]: I0930 14:51:39.359251 4799 generic.go:334] "Generic (PLEG): container finished" podID="8da29a07-04e5-4c03-adf7-f642c86abecd" containerID="4d1c94b5447917d2b8d411a53cab0f977fe5acb19a7c24f759e1c2f732b67bd6" exitCode=0 Sep 30 14:51:39 crc kubenswrapper[4799]: I0930 14:51:39.359344 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" event={"ID":"8da29a07-04e5-4c03-adf7-f642c86abecd","Type":"ContainerDied","Data":"4d1c94b5447917d2b8d411a53cab0f977fe5acb19a7c24f759e1c2f732b67bd6"} Sep 30 14:51:40 crc kubenswrapper[4799]: I0930 14:51:40.829126 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:51:40 crc kubenswrapper[4799]: I0930 14:51:40.916250 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl42h\" (UniqueName: \"kubernetes.io/projected/8da29a07-04e5-4c03-adf7-f642c86abecd-kube-api-access-zl42h\") pod \"8da29a07-04e5-4c03-adf7-f642c86abecd\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " Sep 30 14:51:40 crc kubenswrapper[4799]: I0930 14:51:40.916312 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-ssh-key\") pod \"8da29a07-04e5-4c03-adf7-f642c86abecd\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " Sep 30 14:51:40 crc kubenswrapper[4799]: I0930 14:51:40.917280 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-inventory\") pod \"8da29a07-04e5-4c03-adf7-f642c86abecd\" (UID: \"8da29a07-04e5-4c03-adf7-f642c86abecd\") " Sep 30 14:51:40 crc kubenswrapper[4799]: I0930 14:51:40.922508 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8da29a07-04e5-4c03-adf7-f642c86abecd-kube-api-access-zl42h" (OuterVolumeSpecName: "kube-api-access-zl42h") pod "8da29a07-04e5-4c03-adf7-f642c86abecd" (UID: "8da29a07-04e5-4c03-adf7-f642c86abecd"). InnerVolumeSpecName "kube-api-access-zl42h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:51:40 crc kubenswrapper[4799]: I0930 14:51:40.950955 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8da29a07-04e5-4c03-adf7-f642c86abecd" (UID: "8da29a07-04e5-4c03-adf7-f642c86abecd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:51:40 crc kubenswrapper[4799]: I0930 14:51:40.951601 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-inventory" (OuterVolumeSpecName: "inventory") pod "8da29a07-04e5-4c03-adf7-f642c86abecd" (UID: "8da29a07-04e5-4c03-adf7-f642c86abecd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.019708 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl42h\" (UniqueName: \"kubernetes.io/projected/8da29a07-04e5-4c03-adf7-f642c86abecd-kube-api-access-zl42h\") on node \"crc\" DevicePath \"\"" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.019777 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.019791 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8da29a07-04e5-4c03-adf7-f642c86abecd-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.380940 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" event={"ID":"8da29a07-04e5-4c03-adf7-f642c86abecd","Type":"ContainerDied","Data":"9972227e39af6dd0c015641195ee8f26548c411d7c9c1784b5a0105377875c41"} Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.381305 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9972227e39af6dd0c015641195ee8f26548c411d7c9c1784b5a0105377875c41" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.381023 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-svwrv" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.495844 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt"] Sep 30 14:51:41 crc kubenswrapper[4799]: E0930 14:51:41.496392 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da29a07-04e5-4c03-adf7-f642c86abecd" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.496424 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da29a07-04e5-4c03-adf7-f642c86abecd" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.496682 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8da29a07-04e5-4c03-adf7-f642c86abecd" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.497523 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.500283 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.504334 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.504397 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.505104 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.512430 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt"] Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.534360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.534618 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz5hp\" (UniqueName: \"kubernetes.io/projected/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-kube-api-access-lz5hp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.534657 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.635581 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.635639 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz5hp\" (UniqueName: \"kubernetes.io/projected/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-kube-api-access-lz5hp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.635729 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.641074 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.646638 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.653110 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz5hp\" (UniqueName: \"kubernetes.io/projected/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-kube-api-access-lz5hp\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:41 crc kubenswrapper[4799]: I0930 14:51:41.815220 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:51:42 crc kubenswrapper[4799]: I0930 14:51:42.046219 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-2e33-account-create-t58d8"] Sep 30 14:51:42 crc kubenswrapper[4799]: I0930 14:51:42.061239 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-2e33-account-create-t58d8"] Sep 30 14:51:42 crc kubenswrapper[4799]: I0930 14:51:42.436438 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt"] Sep 30 14:51:42 crc kubenswrapper[4799]: I0930 14:51:42.516839 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d70a7325-993b-4f20-a06c-57cdbeb7d557" path="/var/lib/kubelet/pods/d70a7325-993b-4f20-a06c-57cdbeb7d557/volumes" Sep 30 14:51:43 crc kubenswrapper[4799]: I0930 14:51:43.054968 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8bff-account-create-6pft6"] Sep 30 14:51:43 crc kubenswrapper[4799]: I0930 14:51:43.068397 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8bff-account-create-6pft6"] Sep 30 14:51:43 crc kubenswrapper[4799]: I0930 14:51:43.401371 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" event={"ID":"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c","Type":"ContainerStarted","Data":"6ddb1e0ce2600e413ca1a7cea1c82de71d5778a9931484915b9dfc94ff3ca6e2"} Sep 30 14:51:43 crc kubenswrapper[4799]: I0930 14:51:43.401424 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" event={"ID":"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c","Type":"ContainerStarted","Data":"c7649c0b579fab02fbd15879e5757e23be4716c13cc6291242c5cb8f35eca8bb"} Sep 30 14:51:43 crc kubenswrapper[4799]: I0930 14:51:43.431217 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" podStartSLOduration=2.220502078 podStartE2EDuration="2.431190285s" podCreationTimestamp="2025-09-30 14:51:41 +0000 UTC" firstStartedPulling="2025-09-30 14:51:42.451796465 +0000 UTC m=+1924.535396902" lastFinishedPulling="2025-09-30 14:51:42.662484682 +0000 UTC m=+1924.746085109" observedRunningTime="2025-09-30 14:51:43.424634684 +0000 UTC m=+1925.508235121" watchObservedRunningTime="2025-09-30 14:51:43.431190285 +0000 UTC m=+1925.514790712" Sep 30 14:51:44 crc kubenswrapper[4799]: I0930 14:51:44.526469 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d43c94bd-d76c-4680-bf8d-e2f02fd2279c" path="/var/lib/kubelet/pods/d43c94bd-d76c-4680-bf8d-e2f02fd2279c/volumes" Sep 30 14:51:52 crc kubenswrapper[4799]: I0930 14:51:52.037317 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-2670-account-create-z7z29"] Sep 30 14:51:52 crc kubenswrapper[4799]: I0930 14:51:52.048345 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-2670-account-create-z7z29"] Sep 30 14:51:52 crc kubenswrapper[4799]: I0930 14:51:52.620028 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08dca185-ba7a-439d-96ee-2e5aba753dc5" path="/var/lib/kubelet/pods/08dca185-ba7a-439d-96ee-2e5aba753dc5/volumes" Sep 30 14:52:10 crc kubenswrapper[4799]: I0930 14:52:10.912988 4799 scope.go:117] "RemoveContainer" containerID="2b33610392cc6954114682f78227200f563c71da7b917eb4d9dd69af047008b0" Sep 30 14:52:10 crc kubenswrapper[4799]: I0930 14:52:10.973013 4799 scope.go:117] "RemoveContainer" containerID="58bcbfc7a3b37eee3dee8f9d16542b29635d111d0574dd06c3dfce1c3c268a5f" Sep 30 14:52:11 crc kubenswrapper[4799]: I0930 14:52:11.016358 4799 scope.go:117] "RemoveContainer" containerID="f16a947cdc5b5f1a40d76a7948f221b7b454b1a6dc7cde5ea7f0ece434af8582" Sep 30 14:52:11 crc kubenswrapper[4799]: I0930 14:52:11.087500 4799 scope.go:117] "RemoveContainer" containerID="b6e5eb40df52f01f1e4de26c2853ea416973568133a2d4b7015558a8ab49bb5b" Sep 30 14:52:11 crc kubenswrapper[4799]: I0930 14:52:11.287451 4799 scope.go:117] "RemoveContainer" containerID="45d49dee038e752e605be23615d61aafcd35140e6bc6e1ba7d730f3847ce766b" Sep 30 14:52:11 crc kubenswrapper[4799]: I0930 14:52:11.326720 4799 scope.go:117] "RemoveContainer" containerID="e5b4375d1706fe86bf8d15415458e779a1ce85e7b57531385483f6b52d5e568a" Sep 30 14:52:35 crc kubenswrapper[4799]: I0930 14:52:35.062197 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rtflg"] Sep 30 14:52:35 crc kubenswrapper[4799]: I0930 14:52:35.073925 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rtflg"] Sep 30 14:52:36 crc kubenswrapper[4799]: I0930 14:52:36.515987 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2fe6ff-b58e-4554-8ed1-878001b33af1" path="/var/lib/kubelet/pods/5e2fe6ff-b58e-4554-8ed1-878001b33af1/volumes" Sep 30 14:52:59 crc kubenswrapper[4799]: I0930 14:52:59.650361 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:52:59 crc kubenswrapper[4799]: I0930 14:52:59.651441 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:53:02 crc kubenswrapper[4799]: I0930 14:53:02.199931 4799 generic.go:334] "Generic (PLEG): container finished" podID="2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c" containerID="6ddb1e0ce2600e413ca1a7cea1c82de71d5778a9931484915b9dfc94ff3ca6e2" exitCode=0 Sep 30 14:53:02 crc kubenswrapper[4799]: I0930 14:53:02.200304 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" event={"ID":"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c","Type":"ContainerDied","Data":"6ddb1e0ce2600e413ca1a7cea1c82de71d5778a9931484915b9dfc94ff3ca6e2"} Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.651372 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.687869 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-inventory\") pod \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.688628 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-ssh-key\") pod \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.688817 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz5hp\" (UniqueName: \"kubernetes.io/projected/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-kube-api-access-lz5hp\") pod \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\" (UID: \"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c\") " Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.694598 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-kube-api-access-lz5hp" (OuterVolumeSpecName: "kube-api-access-lz5hp") pod "2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c" (UID: "2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c"). InnerVolumeSpecName "kube-api-access-lz5hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.721779 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-inventory" (OuterVolumeSpecName: "inventory") pod "2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c" (UID: "2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.744889 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c" (UID: "2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.793757 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.793831 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:03 crc kubenswrapper[4799]: I0930 14:53:03.793851 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz5hp\" (UniqueName: \"kubernetes.io/projected/2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c-kube-api-access-lz5hp\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.223524 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" event={"ID":"2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c","Type":"ContainerDied","Data":"c7649c0b579fab02fbd15879e5757e23be4716c13cc6291242c5cb8f35eca8bb"} Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.223781 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7649c0b579fab02fbd15879e5757e23be4716c13cc6291242c5cb8f35eca8bb" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.223717 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.334140 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d"] Sep 30 14:53:04 crc kubenswrapper[4799]: E0930 14:53:04.334785 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.334812 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.335041 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.335908 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.340296 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.340629 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.341155 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.342026 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.361107 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d"] Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.407323 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.407475 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.407546 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl88r\" (UniqueName: \"kubernetes.io/projected/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-kube-api-access-vl88r\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.509951 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.510464 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.510563 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl88r\" (UniqueName: \"kubernetes.io/projected/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-kube-api-access-vl88r\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.517580 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.518052 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.535013 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl88r\" (UniqueName: \"kubernetes.io/projected/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-kube-api-access-vl88r\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9p67d\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:04 crc kubenswrapper[4799]: I0930 14:53:04.654203 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:05 crc kubenswrapper[4799]: I0930 14:53:05.236904 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d"] Sep 30 14:53:06 crc kubenswrapper[4799]: I0930 14:53:06.249216 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" event={"ID":"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99","Type":"ContainerStarted","Data":"a7731bc71f75c86957ca24d7647df9933094386ca6470a777c7414cd2871c1df"} Sep 30 14:53:07 crc kubenswrapper[4799]: I0930 14:53:07.047140 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d6gjx"] Sep 30 14:53:07 crc kubenswrapper[4799]: I0930 14:53:07.058805 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d6gjx"] Sep 30 14:53:07 crc kubenswrapper[4799]: I0930 14:53:07.260168 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" event={"ID":"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99","Type":"ContainerStarted","Data":"6da46ee9402dbf781195fa7d8a4dc234fbff736874578e1e859c41a277689dcb"} Sep 30 14:53:07 crc kubenswrapper[4799]: I0930 14:53:07.283870 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" podStartSLOduration=2.5678454779999997 podStartE2EDuration="3.283844545s" podCreationTimestamp="2025-09-30 14:53:04 +0000 UTC" firstStartedPulling="2025-09-30 14:53:05.246833107 +0000 UTC m=+2007.330433534" lastFinishedPulling="2025-09-30 14:53:05.962832174 +0000 UTC m=+2008.046432601" observedRunningTime="2025-09-30 14:53:07.280871198 +0000 UTC m=+2009.364471645" watchObservedRunningTime="2025-09-30 14:53:07.283844545 +0000 UTC m=+2009.367444972" Sep 30 14:53:08 crc kubenswrapper[4799]: I0930 14:53:08.516232 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e276d3c-eb12-4fc1-9d76-3ffff9307d4d" path="/var/lib/kubelet/pods/4e276d3c-eb12-4fc1-9d76-3ffff9307d4d/volumes" Sep 30 14:53:09 crc kubenswrapper[4799]: I0930 14:53:09.030190 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vwbqw"] Sep 30 14:53:09 crc kubenswrapper[4799]: I0930 14:53:09.037718 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vwbqw"] Sep 30 14:53:10 crc kubenswrapper[4799]: I0930 14:53:10.514523 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdc6f141-7e93-46af-9b59-315ee2bb8025" path="/var/lib/kubelet/pods/bdc6f141-7e93-46af-9b59-315ee2bb8025/volumes" Sep 30 14:53:11 crc kubenswrapper[4799]: I0930 14:53:11.630200 4799 scope.go:117] "RemoveContainer" containerID="d7a54ebcdc008ba4e21e63fb3b06dc2352fdd16160a975d87afa427feef45ae2" Sep 30 14:53:11 crc kubenswrapper[4799]: I0930 14:53:11.680807 4799 scope.go:117] "RemoveContainer" containerID="9bf9a8c5a0d0d1bd96932149a0bf7bcb60228b667a1eab56f23473d1713d2758" Sep 30 14:53:11 crc kubenswrapper[4799]: I0930 14:53:11.736874 4799 scope.go:117] "RemoveContainer" containerID="7164927e50e022176906560d22458e8ff4410f75da5bcbc2fd907d362fb9d9a1" Sep 30 14:53:12 crc kubenswrapper[4799]: I0930 14:53:12.321490 4799 generic.go:334] "Generic (PLEG): container finished" podID="06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99" containerID="6da46ee9402dbf781195fa7d8a4dc234fbff736874578e1e859c41a277689dcb" exitCode=0 Sep 30 14:53:12 crc kubenswrapper[4799]: I0930 14:53:12.321565 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" event={"ID":"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99","Type":"ContainerDied","Data":"6da46ee9402dbf781195fa7d8a4dc234fbff736874578e1e859c41a277689dcb"} Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.787772 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.848737 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-ssh-key\") pod \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.849124 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-inventory\") pod \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.849262 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl88r\" (UniqueName: \"kubernetes.io/projected/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-kube-api-access-vl88r\") pod \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\" (UID: \"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99\") " Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.855300 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-kube-api-access-vl88r" (OuterVolumeSpecName: "kube-api-access-vl88r") pod "06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99" (UID: "06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99"). InnerVolumeSpecName "kube-api-access-vl88r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.884627 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-inventory" (OuterVolumeSpecName: "inventory") pod "06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99" (UID: "06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.889173 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99" (UID: "06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.952750 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.952790 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:13 crc kubenswrapper[4799]: I0930 14:53:13.952803 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl88r\" (UniqueName: \"kubernetes.io/projected/06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99-kube-api-access-vl88r\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.343776 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" event={"ID":"06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99","Type":"ContainerDied","Data":"a7731bc71f75c86957ca24d7647df9933094386ca6470a777c7414cd2871c1df"} Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.343818 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9p67d" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.343836 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7731bc71f75c86957ca24d7647df9933094386ca6470a777c7414cd2871c1df" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.529373 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz"] Sep 30 14:53:14 crc kubenswrapper[4799]: E0930 14:53:14.529871 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.529963 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.530255 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.531074 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.545914 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.546038 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.546122 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz"] Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.546459 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.549182 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.569254 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.569756 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dlk7\" (UniqueName: \"kubernetes.io/projected/9a20125e-0fe9-4826-b17d-f7b61c73bac3-kube-api-access-8dlk7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.569957 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.672344 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.672507 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dlk7\" (UniqueName: \"kubernetes.io/projected/9a20125e-0fe9-4826-b17d-f7b61c73bac3-kube-api-access-8dlk7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.672623 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.678643 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.680030 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.695583 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dlk7\" (UniqueName: \"kubernetes.io/projected/9a20125e-0fe9-4826-b17d-f7b61c73bac3-kube-api-access-8dlk7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6f7fz\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:14 crc kubenswrapper[4799]: I0930 14:53:14.872414 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:15 crc kubenswrapper[4799]: I0930 14:53:15.425921 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz"] Sep 30 14:53:16 crc kubenswrapper[4799]: I0930 14:53:16.364013 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" event={"ID":"9a20125e-0fe9-4826-b17d-f7b61c73bac3","Type":"ContainerStarted","Data":"1d2969419b93645d18e821ec8cc902f86b49833e3a7a9bf5b4de17e3d764dfc4"} Sep 30 14:53:16 crc kubenswrapper[4799]: I0930 14:53:16.364334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" event={"ID":"9a20125e-0fe9-4826-b17d-f7b61c73bac3","Type":"ContainerStarted","Data":"ba0e04a55586eb520cb7ec7e3434b1f81ddfa53b0e0e185ff68d220994853a53"} Sep 30 14:53:16 crc kubenswrapper[4799]: I0930 14:53:16.387131 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" podStartSLOduration=2.186108728 podStartE2EDuration="2.387105313s" podCreationTimestamp="2025-09-30 14:53:14 +0000 UTC" firstStartedPulling="2025-09-30 14:53:15.439637273 +0000 UTC m=+2017.523237700" lastFinishedPulling="2025-09-30 14:53:15.640633868 +0000 UTC m=+2017.724234285" observedRunningTime="2025-09-30 14:53:16.385361932 +0000 UTC m=+2018.468962389" watchObservedRunningTime="2025-09-30 14:53:16.387105313 +0000 UTC m=+2018.470705740" Sep 30 14:53:29 crc kubenswrapper[4799]: I0930 14:53:29.648494 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:53:29 crc kubenswrapper[4799]: I0930 14:53:29.649087 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:53:57 crc kubenswrapper[4799]: I0930 14:53:57.781272 4799 generic.go:334] "Generic (PLEG): container finished" podID="9a20125e-0fe9-4826-b17d-f7b61c73bac3" containerID="1d2969419b93645d18e821ec8cc902f86b49833e3a7a9bf5b4de17e3d764dfc4" exitCode=0 Sep 30 14:53:57 crc kubenswrapper[4799]: I0930 14:53:57.781349 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" event={"ID":"9a20125e-0fe9-4826-b17d-f7b61c73bac3","Type":"ContainerDied","Data":"1d2969419b93645d18e821ec8cc902f86b49833e3a7a9bf5b4de17e3d764dfc4"} Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.193984 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.251835 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dlk7\" (UniqueName: \"kubernetes.io/projected/9a20125e-0fe9-4826-b17d-f7b61c73bac3-kube-api-access-8dlk7\") pod \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.252441 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-inventory\") pod \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.252513 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-ssh-key\") pod \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\" (UID: \"9a20125e-0fe9-4826-b17d-f7b61c73bac3\") " Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.272138 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a20125e-0fe9-4826-b17d-f7b61c73bac3-kube-api-access-8dlk7" (OuterVolumeSpecName: "kube-api-access-8dlk7") pod "9a20125e-0fe9-4826-b17d-f7b61c73bac3" (UID: "9a20125e-0fe9-4826-b17d-f7b61c73bac3"). InnerVolumeSpecName "kube-api-access-8dlk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.286821 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9a20125e-0fe9-4826-b17d-f7b61c73bac3" (UID: "9a20125e-0fe9-4826-b17d-f7b61c73bac3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.288937 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-inventory" (OuterVolumeSpecName: "inventory") pod "9a20125e-0fe9-4826-b17d-f7b61c73bac3" (UID: "9a20125e-0fe9-4826-b17d-f7b61c73bac3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.356066 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dlk7\" (UniqueName: \"kubernetes.io/projected/9a20125e-0fe9-4826-b17d-f7b61c73bac3-kube-api-access-8dlk7\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.356116 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.356131 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a20125e-0fe9-4826-b17d-f7b61c73bac3-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.648558 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.648899 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.649065 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.649818 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"116aa297eadbb7887d7d899e95966cb2028cb7fa70dc029dd1713cd3c032c2a4"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.649970 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://116aa297eadbb7887d7d899e95966cb2028cb7fa70dc029dd1713cd3c032c2a4" gracePeriod=600 Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.825150 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="116aa297eadbb7887d7d899e95966cb2028cb7fa70dc029dd1713cd3c032c2a4" exitCode=0 Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.829794 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"116aa297eadbb7887d7d899e95966cb2028cb7fa70dc029dd1713cd3c032c2a4"} Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.829869 4799 scope.go:117] "RemoveContainer" containerID="511c9071501850a308a7c611917cbe8780f98baae5e949da894bb14c96b10fd1" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.842261 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" event={"ID":"9a20125e-0fe9-4826-b17d-f7b61c73bac3","Type":"ContainerDied","Data":"ba0e04a55586eb520cb7ec7e3434b1f81ddfa53b0e0e185ff68d220994853a53"} Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.842634 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba0e04a55586eb520cb7ec7e3434b1f81ddfa53b0e0e185ff68d220994853a53" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.842806 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6f7fz" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.939665 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t"] Sep 30 14:53:59 crc kubenswrapper[4799]: E0930 14:53:59.940569 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a20125e-0fe9-4826-b17d-f7b61c73bac3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.940708 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a20125e-0fe9-4826-b17d-f7b61c73bac3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.941039 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a20125e-0fe9-4826-b17d-f7b61c73bac3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.941827 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.947771 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.948125 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.948242 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.948318 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.969962 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t"] Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.970833 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.971010 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l792b\" (UniqueName: \"kubernetes.io/projected/e2cb3e85-dce2-4274-9472-3d5b10ba4358-kube-api-access-l792b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:53:59 crc kubenswrapper[4799]: I0930 14:53:59.971250 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.043742 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-8xfj5"] Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.058018 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-8xfj5"] Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.073399 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.073538 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.073707 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l792b\" (UniqueName: \"kubernetes.io/projected/e2cb3e85-dce2-4274-9472-3d5b10ba4358-kube-api-access-l792b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.080383 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.084198 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.097699 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l792b\" (UniqueName: \"kubernetes.io/projected/e2cb3e85-dce2-4274-9472-3d5b10ba4358-kube-api-access-l792b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llf6t\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.276321 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.517839 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0231f24-294e-400b-99ac-2bc5c5e97e25" path="/var/lib/kubelet/pods/e0231f24-294e-400b-99ac-2bc5c5e97e25/volumes" Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.858609 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t"] Sep 30 14:54:00 crc kubenswrapper[4799]: I0930 14:54:00.860663 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976"} Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.856714 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kbmkz"] Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.859204 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.883686 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbmkz"] Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.885258 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" event={"ID":"e2cb3e85-dce2-4274-9472-3d5b10ba4358","Type":"ContainerStarted","Data":"d48664bf4e3b6abd46ba3649085f8ffc3f0144afe213da0f39a0df4e0bf7e789"} Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.885294 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" event={"ID":"e2cb3e85-dce2-4274-9472-3d5b10ba4358","Type":"ContainerStarted","Data":"71bf64fd3bf88e0e027ef821dd0880b12bd13afbfa907f673669d7aa624e5a4c"} Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.919990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-utilities\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.920089 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-catalog-content\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.920278 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvtpq\" (UniqueName: \"kubernetes.io/projected/47526621-4977-4eb3-9402-80a99cb1104c-kube-api-access-cvtpq\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:01 crc kubenswrapper[4799]: I0930 14:54:01.928174 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" podStartSLOduration=2.648218267 podStartE2EDuration="2.928139711s" podCreationTimestamp="2025-09-30 14:53:59 +0000 UTC" firstStartedPulling="2025-09-30 14:54:00.869625896 +0000 UTC m=+2062.953226323" lastFinishedPulling="2025-09-30 14:54:01.14954734 +0000 UTC m=+2063.233147767" observedRunningTime="2025-09-30 14:54:01.912800154 +0000 UTC m=+2063.996400591" watchObservedRunningTime="2025-09-30 14:54:01.928139711 +0000 UTC m=+2064.011740138" Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.021825 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvtpq\" (UniqueName: \"kubernetes.io/projected/47526621-4977-4eb3-9402-80a99cb1104c-kube-api-access-cvtpq\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.021921 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-utilities\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.021973 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-catalog-content\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.022428 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-catalog-content\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.023146 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-utilities\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.062861 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvtpq\" (UniqueName: \"kubernetes.io/projected/47526621-4977-4eb3-9402-80a99cb1104c-kube-api-access-cvtpq\") pod \"community-operators-kbmkz\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.196142 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.792939 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbmkz"] Sep 30 14:54:02 crc kubenswrapper[4799]: W0930 14:54:02.820931 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47526621_4977_4eb3_9402_80a99cb1104c.slice/crio-413c9479f32d16875c649af797f6c125d1b5441a56884b4eb478a122514469ae WatchSource:0}: Error finding container 413c9479f32d16875c649af797f6c125d1b5441a56884b4eb478a122514469ae: Status 404 returned error can't find the container with id 413c9479f32d16875c649af797f6c125d1b5441a56884b4eb478a122514469ae Sep 30 14:54:02 crc kubenswrapper[4799]: I0930 14:54:02.902058 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbmkz" event={"ID":"47526621-4977-4eb3-9402-80a99cb1104c","Type":"ContainerStarted","Data":"413c9479f32d16875c649af797f6c125d1b5441a56884b4eb478a122514469ae"} Sep 30 14:54:03 crc kubenswrapper[4799]: I0930 14:54:03.913636 4799 generic.go:334] "Generic (PLEG): container finished" podID="47526621-4977-4eb3-9402-80a99cb1104c" containerID="ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e" exitCode=0 Sep 30 14:54:03 crc kubenswrapper[4799]: I0930 14:54:03.913750 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbmkz" event={"ID":"47526621-4977-4eb3-9402-80a99cb1104c","Type":"ContainerDied","Data":"ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e"} Sep 30 14:54:04 crc kubenswrapper[4799]: I0930 14:54:04.927367 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbmkz" event={"ID":"47526621-4977-4eb3-9402-80a99cb1104c","Type":"ContainerStarted","Data":"79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268"} Sep 30 14:54:06 crc kubenswrapper[4799]: I0930 14:54:06.950791 4799 generic.go:334] "Generic (PLEG): container finished" podID="47526621-4977-4eb3-9402-80a99cb1104c" containerID="79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268" exitCode=0 Sep 30 14:54:06 crc kubenswrapper[4799]: I0930 14:54:06.950854 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbmkz" event={"ID":"47526621-4977-4eb3-9402-80a99cb1104c","Type":"ContainerDied","Data":"79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268"} Sep 30 14:54:07 crc kubenswrapper[4799]: I0930 14:54:07.963488 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbmkz" event={"ID":"47526621-4977-4eb3-9402-80a99cb1104c","Type":"ContainerStarted","Data":"b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed"} Sep 30 14:54:08 crc kubenswrapper[4799]: I0930 14:54:08.017623 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kbmkz" podStartSLOduration=3.536482392 podStartE2EDuration="7.017595397s" podCreationTimestamp="2025-09-30 14:54:01 +0000 UTC" firstStartedPulling="2025-09-30 14:54:03.916684686 +0000 UTC m=+2066.000285113" lastFinishedPulling="2025-09-30 14:54:07.397797691 +0000 UTC m=+2069.481398118" observedRunningTime="2025-09-30 14:54:08.016287599 +0000 UTC m=+2070.099888046" watchObservedRunningTime="2025-09-30 14:54:08.017595397 +0000 UTC m=+2070.101195824" Sep 30 14:54:11 crc kubenswrapper[4799]: I0930 14:54:11.870236 4799 scope.go:117] "RemoveContainer" containerID="7942a311c59954911c37813fa7db634c655fb3fae8a1f3706f31a3c67322561b" Sep 30 14:54:12 crc kubenswrapper[4799]: I0930 14:54:12.195192 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:12 crc kubenswrapper[4799]: I0930 14:54:12.195589 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:12 crc kubenswrapper[4799]: I0930 14:54:12.256219 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:13 crc kubenswrapper[4799]: I0930 14:54:13.054966 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:13 crc kubenswrapper[4799]: I0930 14:54:13.110777 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kbmkz"] Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.026637 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kbmkz" podUID="47526621-4977-4eb3-9402-80a99cb1104c" containerName="registry-server" containerID="cri-o://b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed" gracePeriod=2 Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.524750 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.545374 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-utilities\") pod \"47526621-4977-4eb3-9402-80a99cb1104c\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.545484 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvtpq\" (UniqueName: \"kubernetes.io/projected/47526621-4977-4eb3-9402-80a99cb1104c-kube-api-access-cvtpq\") pod \"47526621-4977-4eb3-9402-80a99cb1104c\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.545675 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-catalog-content\") pod \"47526621-4977-4eb3-9402-80a99cb1104c\" (UID: \"47526621-4977-4eb3-9402-80a99cb1104c\") " Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.547103 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-utilities" (OuterVolumeSpecName: "utilities") pod "47526621-4977-4eb3-9402-80a99cb1104c" (UID: "47526621-4977-4eb3-9402-80a99cb1104c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.576533 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.592982 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47526621-4977-4eb3-9402-80a99cb1104c-kube-api-access-cvtpq" (OuterVolumeSpecName: "kube-api-access-cvtpq") pod "47526621-4977-4eb3-9402-80a99cb1104c" (UID: "47526621-4977-4eb3-9402-80a99cb1104c"). InnerVolumeSpecName "kube-api-access-cvtpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.607601 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47526621-4977-4eb3-9402-80a99cb1104c" (UID: "47526621-4977-4eb3-9402-80a99cb1104c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.678155 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvtpq\" (UniqueName: \"kubernetes.io/projected/47526621-4977-4eb3-9402-80a99cb1104c-kube-api-access-cvtpq\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:15 crc kubenswrapper[4799]: I0930 14:54:15.678232 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47526621-4977-4eb3-9402-80a99cb1104c-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.042028 4799 generic.go:334] "Generic (PLEG): container finished" podID="47526621-4977-4eb3-9402-80a99cb1104c" containerID="b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed" exitCode=0 Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.042105 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbmkz" event={"ID":"47526621-4977-4eb3-9402-80a99cb1104c","Type":"ContainerDied","Data":"b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed"} Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.042145 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbmkz" event={"ID":"47526621-4977-4eb3-9402-80a99cb1104c","Type":"ContainerDied","Data":"413c9479f32d16875c649af797f6c125d1b5441a56884b4eb478a122514469ae"} Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.042169 4799 scope.go:117] "RemoveContainer" containerID="b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.042175 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbmkz" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.075827 4799 scope.go:117] "RemoveContainer" containerID="79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.090355 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kbmkz"] Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.100516 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kbmkz"] Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.114104 4799 scope.go:117] "RemoveContainer" containerID="ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.150744 4799 scope.go:117] "RemoveContainer" containerID="b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed" Sep 30 14:54:16 crc kubenswrapper[4799]: E0930 14:54:16.151620 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed\": container with ID starting with b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed not found: ID does not exist" containerID="b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.151707 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed"} err="failed to get container status \"b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed\": rpc error: code = NotFound desc = could not find container \"b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed\": container with ID starting with b606cc07ee377a7c835ae4e708a770b03d91b35aefe58758ad4d730943662eed not found: ID does not exist" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.151747 4799 scope.go:117] "RemoveContainer" containerID="79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268" Sep 30 14:54:16 crc kubenswrapper[4799]: E0930 14:54:16.152078 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268\": container with ID starting with 79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268 not found: ID does not exist" containerID="79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.152122 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268"} err="failed to get container status \"79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268\": rpc error: code = NotFound desc = could not find container \"79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268\": container with ID starting with 79cfec16e84baa74c82c046e1bb9dd7e9012a052769e2789cbab74e90aa5c268 not found: ID does not exist" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.152161 4799 scope.go:117] "RemoveContainer" containerID="ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e" Sep 30 14:54:16 crc kubenswrapper[4799]: E0930 14:54:16.152464 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e\": container with ID starting with ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e not found: ID does not exist" containerID="ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.152570 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e"} err="failed to get container status \"ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e\": rpc error: code = NotFound desc = could not find container \"ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e\": container with ID starting with ca4955d1b0af7c9e5ef879154d93f3745dbb269719f9634e0d96489f39b9193e not found: ID does not exist" Sep 30 14:54:16 crc kubenswrapper[4799]: I0930 14:54:16.517467 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47526621-4977-4eb3-9402-80a99cb1104c" path="/var/lib/kubelet/pods/47526621-4977-4eb3-9402-80a99cb1104c/volumes" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.314480 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wgxlp"] Sep 30 14:54:34 crc kubenswrapper[4799]: E0930 14:54:34.315722 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47526621-4977-4eb3-9402-80a99cb1104c" containerName="extract-content" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.315747 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="47526621-4977-4eb3-9402-80a99cb1104c" containerName="extract-content" Sep 30 14:54:34 crc kubenswrapper[4799]: E0930 14:54:34.315768 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47526621-4977-4eb3-9402-80a99cb1104c" containerName="registry-server" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.315778 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="47526621-4977-4eb3-9402-80a99cb1104c" containerName="registry-server" Sep 30 14:54:34 crc kubenswrapper[4799]: E0930 14:54:34.315833 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47526621-4977-4eb3-9402-80a99cb1104c" containerName="extract-utilities" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.315842 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="47526621-4977-4eb3-9402-80a99cb1104c" containerName="extract-utilities" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.316066 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="47526621-4977-4eb3-9402-80a99cb1104c" containerName="registry-server" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.318089 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.337268 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgxlp"] Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.355858 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcszx\" (UniqueName: \"kubernetes.io/projected/8277152e-d298-44b1-8dc8-127b6a3349f2-kube-api-access-jcszx\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.356041 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-catalog-content\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.356109 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-utilities\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.458013 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-catalog-content\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.458100 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-utilities\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.458156 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcszx\" (UniqueName: \"kubernetes.io/projected/8277152e-d298-44b1-8dc8-127b6a3349f2-kube-api-access-jcszx\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.458816 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-catalog-content\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.458831 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-utilities\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.495110 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcszx\" (UniqueName: \"kubernetes.io/projected/8277152e-d298-44b1-8dc8-127b6a3349f2-kube-api-access-jcszx\") pod \"redhat-marketplace-wgxlp\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:34 crc kubenswrapper[4799]: I0930 14:54:34.639811 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:35 crc kubenswrapper[4799]: I0930 14:54:35.154167 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgxlp"] Sep 30 14:54:35 crc kubenswrapper[4799]: I0930 14:54:35.209249 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgxlp" event={"ID":"8277152e-d298-44b1-8dc8-127b6a3349f2","Type":"ContainerStarted","Data":"0ba8f055514d30a968896cb619485961795bf1ac275d9918c3dfd42fdc82e44b"} Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.220003 4799 generic.go:334] "Generic (PLEG): container finished" podID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerID="993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b" exitCode=0 Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.220128 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgxlp" event={"ID":"8277152e-d298-44b1-8dc8-127b6a3349f2","Type":"ContainerDied","Data":"993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b"} Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.726734 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z2fhg"] Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.729530 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.749644 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2fhg"] Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.804425 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vlq2\" (UniqueName: \"kubernetes.io/projected/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-kube-api-access-4vlq2\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.804557 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-catalog-content\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.804751 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-utilities\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.906858 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vlq2\" (UniqueName: \"kubernetes.io/projected/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-kube-api-access-4vlq2\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.906949 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-catalog-content\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.907074 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-utilities\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.907690 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-utilities\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.908428 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-catalog-content\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:36 crc kubenswrapper[4799]: I0930 14:54:36.936466 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vlq2\" (UniqueName: \"kubernetes.io/projected/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-kube-api-access-4vlq2\") pod \"redhat-operators-z2fhg\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:37 crc kubenswrapper[4799]: I0930 14:54:37.063145 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:37 crc kubenswrapper[4799]: I0930 14:54:37.583412 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2fhg"] Sep 30 14:54:37 crc kubenswrapper[4799]: W0930 14:54:37.606074 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cbb077e_8e14_41fd_bc6c_3c47751ded5a.slice/crio-cb2d60914db493e68af106ed88cd7cf9d1ba3a6ac325da0e0a0c833b0f2235c3 WatchSource:0}: Error finding container cb2d60914db493e68af106ed88cd7cf9d1ba3a6ac325da0e0a0c833b0f2235c3: Status 404 returned error can't find the container with id cb2d60914db493e68af106ed88cd7cf9d1ba3a6ac325da0e0a0c833b0f2235c3 Sep 30 14:54:38 crc kubenswrapper[4799]: I0930 14:54:38.245977 4799 generic.go:334] "Generic (PLEG): container finished" podID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerID="c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111" exitCode=0 Sep 30 14:54:38 crc kubenswrapper[4799]: I0930 14:54:38.246132 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2fhg" event={"ID":"5cbb077e-8e14-41fd-bc6c-3c47751ded5a","Type":"ContainerDied","Data":"c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111"} Sep 30 14:54:38 crc kubenswrapper[4799]: I0930 14:54:38.246396 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2fhg" event={"ID":"5cbb077e-8e14-41fd-bc6c-3c47751ded5a","Type":"ContainerStarted","Data":"cb2d60914db493e68af106ed88cd7cf9d1ba3a6ac325da0e0a0c833b0f2235c3"} Sep 30 14:54:38 crc kubenswrapper[4799]: I0930 14:54:38.248743 4799 generic.go:334] "Generic (PLEG): container finished" podID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerID="e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861" exitCode=0 Sep 30 14:54:38 crc kubenswrapper[4799]: I0930 14:54:38.248770 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgxlp" event={"ID":"8277152e-d298-44b1-8dc8-127b6a3349f2","Type":"ContainerDied","Data":"e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861"} Sep 30 14:54:39 crc kubenswrapper[4799]: I0930 14:54:39.263942 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgxlp" event={"ID":"8277152e-d298-44b1-8dc8-127b6a3349f2","Type":"ContainerStarted","Data":"bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb"} Sep 30 14:54:39 crc kubenswrapper[4799]: I0930 14:54:39.298498 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wgxlp" podStartSLOduration=2.724347142 podStartE2EDuration="5.298470667s" podCreationTimestamp="2025-09-30 14:54:34 +0000 UTC" firstStartedPulling="2025-09-30 14:54:36.225548563 +0000 UTC m=+2098.309148990" lastFinishedPulling="2025-09-30 14:54:38.799672098 +0000 UTC m=+2100.883272515" observedRunningTime="2025-09-30 14:54:39.29169037 +0000 UTC m=+2101.375290817" watchObservedRunningTime="2025-09-30 14:54:39.298470667 +0000 UTC m=+2101.382071094" Sep 30 14:54:40 crc kubenswrapper[4799]: I0930 14:54:40.295288 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2fhg" event={"ID":"5cbb077e-8e14-41fd-bc6c-3c47751ded5a","Type":"ContainerStarted","Data":"45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872"} Sep 30 14:54:44 crc kubenswrapper[4799]: I0930 14:54:44.333616 4799 generic.go:334] "Generic (PLEG): container finished" podID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerID="45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872" exitCode=0 Sep 30 14:54:44 crc kubenswrapper[4799]: I0930 14:54:44.333829 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2fhg" event={"ID":"5cbb077e-8e14-41fd-bc6c-3c47751ded5a","Type":"ContainerDied","Data":"45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872"} Sep 30 14:54:44 crc kubenswrapper[4799]: I0930 14:54:44.640369 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:44 crc kubenswrapper[4799]: I0930 14:54:44.640444 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:44 crc kubenswrapper[4799]: I0930 14:54:44.690498 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:45 crc kubenswrapper[4799]: I0930 14:54:45.349095 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2fhg" event={"ID":"5cbb077e-8e14-41fd-bc6c-3c47751ded5a","Type":"ContainerStarted","Data":"891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9"} Sep 30 14:54:45 crc kubenswrapper[4799]: I0930 14:54:45.379688 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z2fhg" podStartSLOduration=2.8895351959999998 podStartE2EDuration="9.379635264s" podCreationTimestamp="2025-09-30 14:54:36 +0000 UTC" firstStartedPulling="2025-09-30 14:54:38.247772211 +0000 UTC m=+2100.331372638" lastFinishedPulling="2025-09-30 14:54:44.737872279 +0000 UTC m=+2106.821472706" observedRunningTime="2025-09-30 14:54:45.370817977 +0000 UTC m=+2107.454418414" watchObservedRunningTime="2025-09-30 14:54:45.379635264 +0000 UTC m=+2107.463235691" Sep 30 14:54:45 crc kubenswrapper[4799]: I0930 14:54:45.407062 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:47 crc kubenswrapper[4799]: I0930 14:54:47.063868 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:47 crc kubenswrapper[4799]: I0930 14:54:47.064764 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:47 crc kubenswrapper[4799]: I0930 14:54:47.105352 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgxlp"] Sep 30 14:54:48 crc kubenswrapper[4799]: I0930 14:54:48.125112 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z2fhg" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="registry-server" probeResult="failure" output=< Sep 30 14:54:48 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 14:54:48 crc kubenswrapper[4799]: > Sep 30 14:54:48 crc kubenswrapper[4799]: I0930 14:54:48.379128 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wgxlp" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerName="registry-server" containerID="cri-o://bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb" gracePeriod=2 Sep 30 14:54:48 crc kubenswrapper[4799]: I0930 14:54:48.900217 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.069089 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcszx\" (UniqueName: \"kubernetes.io/projected/8277152e-d298-44b1-8dc8-127b6a3349f2-kube-api-access-jcszx\") pod \"8277152e-d298-44b1-8dc8-127b6a3349f2\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.069258 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-catalog-content\") pod \"8277152e-d298-44b1-8dc8-127b6a3349f2\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.069518 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-utilities\") pod \"8277152e-d298-44b1-8dc8-127b6a3349f2\" (UID: \"8277152e-d298-44b1-8dc8-127b6a3349f2\") " Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.069976 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-utilities" (OuterVolumeSpecName: "utilities") pod "8277152e-d298-44b1-8dc8-127b6a3349f2" (UID: "8277152e-d298-44b1-8dc8-127b6a3349f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.070437 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.076882 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8277152e-d298-44b1-8dc8-127b6a3349f2-kube-api-access-jcszx" (OuterVolumeSpecName: "kube-api-access-jcszx") pod "8277152e-d298-44b1-8dc8-127b6a3349f2" (UID: "8277152e-d298-44b1-8dc8-127b6a3349f2"). InnerVolumeSpecName "kube-api-access-jcszx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.082008 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8277152e-d298-44b1-8dc8-127b6a3349f2" (UID: "8277152e-d298-44b1-8dc8-127b6a3349f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.172273 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcszx\" (UniqueName: \"kubernetes.io/projected/8277152e-d298-44b1-8dc8-127b6a3349f2-kube-api-access-jcszx\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.172327 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8277152e-d298-44b1-8dc8-127b6a3349f2-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.392496 4799 generic.go:334] "Generic (PLEG): container finished" podID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerID="bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb" exitCode=0 Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.392554 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgxlp" event={"ID":"8277152e-d298-44b1-8dc8-127b6a3349f2","Type":"ContainerDied","Data":"bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb"} Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.392596 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgxlp" event={"ID":"8277152e-d298-44b1-8dc8-127b6a3349f2","Type":"ContainerDied","Data":"0ba8f055514d30a968896cb619485961795bf1ac275d9918c3dfd42fdc82e44b"} Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.392618 4799 scope.go:117] "RemoveContainer" containerID="bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.392811 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgxlp" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.438794 4799 scope.go:117] "RemoveContainer" containerID="e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.444822 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgxlp"] Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.453602 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgxlp"] Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.471053 4799 scope.go:117] "RemoveContainer" containerID="993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.562887 4799 scope.go:117] "RemoveContainer" containerID="bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb" Sep 30 14:54:49 crc kubenswrapper[4799]: E0930 14:54:49.564396 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb\": container with ID starting with bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb not found: ID does not exist" containerID="bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.564484 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb"} err="failed to get container status \"bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb\": rpc error: code = NotFound desc = could not find container \"bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb\": container with ID starting with bca61d590c0e852663f3908cf4bcfa8079aaa79eea16c4656c57d89a62cff5fb not found: ID does not exist" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.564527 4799 scope.go:117] "RemoveContainer" containerID="e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861" Sep 30 14:54:49 crc kubenswrapper[4799]: E0930 14:54:49.565907 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861\": container with ID starting with e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861 not found: ID does not exist" containerID="e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.565946 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861"} err="failed to get container status \"e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861\": rpc error: code = NotFound desc = could not find container \"e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861\": container with ID starting with e34d8372958887c175a7f801351677c6621714bc1aaf0de2a1308214827af861 not found: ID does not exist" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.565974 4799 scope.go:117] "RemoveContainer" containerID="993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b" Sep 30 14:54:49 crc kubenswrapper[4799]: E0930 14:54:49.566740 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b\": container with ID starting with 993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b not found: ID does not exist" containerID="993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b" Sep 30 14:54:49 crc kubenswrapper[4799]: I0930 14:54:49.566762 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b"} err="failed to get container status \"993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b\": rpc error: code = NotFound desc = could not find container \"993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b\": container with ID starting with 993c351be3beb770dca2bb2ab297bbecdfdb90c505b68686c3f745c867960c8b not found: ID does not exist" Sep 30 14:54:50 crc kubenswrapper[4799]: I0930 14:54:50.521548 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" path="/var/lib/kubelet/pods/8277152e-d298-44b1-8dc8-127b6a3349f2/volumes" Sep 30 14:54:57 crc kubenswrapper[4799]: I0930 14:54:57.116435 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:57 crc kubenswrapper[4799]: I0930 14:54:57.173777 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:57 crc kubenswrapper[4799]: I0930 14:54:57.357307 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2fhg"] Sep 30 14:54:58 crc kubenswrapper[4799]: I0930 14:54:58.474808 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z2fhg" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="registry-server" containerID="cri-o://891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9" gracePeriod=2 Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.003038 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.007478 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vlq2\" (UniqueName: \"kubernetes.io/projected/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-kube-api-access-4vlq2\") pod \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.022134 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-kube-api-access-4vlq2" (OuterVolumeSpecName: "kube-api-access-4vlq2") pod "5cbb077e-8e14-41fd-bc6c-3c47751ded5a" (UID: "5cbb077e-8e14-41fd-bc6c-3c47751ded5a"). InnerVolumeSpecName "kube-api-access-4vlq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.109452 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-utilities\") pod \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.109866 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-catalog-content\") pod \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\" (UID: \"5cbb077e-8e14-41fd-bc6c-3c47751ded5a\") " Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.110545 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vlq2\" (UniqueName: \"kubernetes.io/projected/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-kube-api-access-4vlq2\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.110669 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-utilities" (OuterVolumeSpecName: "utilities") pod "5cbb077e-8e14-41fd-bc6c-3c47751ded5a" (UID: "5cbb077e-8e14-41fd-bc6c-3c47751ded5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.200829 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cbb077e-8e14-41fd-bc6c-3c47751ded5a" (UID: "5cbb077e-8e14-41fd-bc6c-3c47751ded5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.212189 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.212226 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cbb077e-8e14-41fd-bc6c-3c47751ded5a-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.488628 4799 generic.go:334] "Generic (PLEG): container finished" podID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerID="891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9" exitCode=0 Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.488706 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2fhg" event={"ID":"5cbb077e-8e14-41fd-bc6c-3c47751ded5a","Type":"ContainerDied","Data":"891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9"} Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.488751 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2fhg" event={"ID":"5cbb077e-8e14-41fd-bc6c-3c47751ded5a","Type":"ContainerDied","Data":"cb2d60914db493e68af106ed88cd7cf9d1ba3a6ac325da0e0a0c833b0f2235c3"} Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.488754 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2fhg" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.488776 4799 scope.go:117] "RemoveContainer" containerID="891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.527896 4799 scope.go:117] "RemoveContainer" containerID="45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.529586 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2fhg"] Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.551643 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z2fhg"] Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.591693 4799 scope.go:117] "RemoveContainer" containerID="c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.616334 4799 scope.go:117] "RemoveContainer" containerID="891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9" Sep 30 14:54:59 crc kubenswrapper[4799]: E0930 14:54:59.617251 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9\": container with ID starting with 891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9 not found: ID does not exist" containerID="891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.617330 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9"} err="failed to get container status \"891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9\": rpc error: code = NotFound desc = could not find container \"891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9\": container with ID starting with 891657ac6daf4949386b1e35ee27d2d8c65bf60d304ed50e3ec0f49adbc847a9 not found: ID does not exist" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.617381 4799 scope.go:117] "RemoveContainer" containerID="45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872" Sep 30 14:54:59 crc kubenswrapper[4799]: E0930 14:54:59.617843 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872\": container with ID starting with 45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872 not found: ID does not exist" containerID="45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.617879 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872"} err="failed to get container status \"45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872\": rpc error: code = NotFound desc = could not find container \"45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872\": container with ID starting with 45fc6d81c979a84c8ff567733c84be1ae55f8f29ab302b11926dcd061808c872 not found: ID does not exist" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.617904 4799 scope.go:117] "RemoveContainer" containerID="c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111" Sep 30 14:54:59 crc kubenswrapper[4799]: E0930 14:54:59.618218 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111\": container with ID starting with c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111 not found: ID does not exist" containerID="c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111" Sep 30 14:54:59 crc kubenswrapper[4799]: I0930 14:54:59.618263 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111"} err="failed to get container status \"c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111\": rpc error: code = NotFound desc = could not find container \"c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111\": container with ID starting with c4c0e4856c91affaf660aa1fcc7bfc5b72b0fd865fc0b351138b5aa34a61b111 not found: ID does not exist" Sep 30 14:55:00 crc kubenswrapper[4799]: I0930 14:55:00.499363 4799 generic.go:334] "Generic (PLEG): container finished" podID="e2cb3e85-dce2-4274-9472-3d5b10ba4358" containerID="d48664bf4e3b6abd46ba3649085f8ffc3f0144afe213da0f39a0df4e0bf7e789" exitCode=2 Sep 30 14:55:00 crc kubenswrapper[4799]: I0930 14:55:00.499378 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" event={"ID":"e2cb3e85-dce2-4274-9472-3d5b10ba4358","Type":"ContainerDied","Data":"d48664bf4e3b6abd46ba3649085f8ffc3f0144afe213da0f39a0df4e0bf7e789"} Sep 30 14:55:00 crc kubenswrapper[4799]: I0930 14:55:00.516256 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" path="/var/lib/kubelet/pods/5cbb077e-8e14-41fd-bc6c-3c47751ded5a/volumes" Sep 30 14:55:01 crc kubenswrapper[4799]: I0930 14:55:01.917005 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.078755 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-inventory\") pod \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.078882 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l792b\" (UniqueName: \"kubernetes.io/projected/e2cb3e85-dce2-4274-9472-3d5b10ba4358-kube-api-access-l792b\") pod \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.078955 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-ssh-key\") pod \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\" (UID: \"e2cb3e85-dce2-4274-9472-3d5b10ba4358\") " Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.085318 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2cb3e85-dce2-4274-9472-3d5b10ba4358-kube-api-access-l792b" (OuterVolumeSpecName: "kube-api-access-l792b") pod "e2cb3e85-dce2-4274-9472-3d5b10ba4358" (UID: "e2cb3e85-dce2-4274-9472-3d5b10ba4358"). InnerVolumeSpecName "kube-api-access-l792b". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.116467 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-inventory" (OuterVolumeSpecName: "inventory") pod "e2cb3e85-dce2-4274-9472-3d5b10ba4358" (UID: "e2cb3e85-dce2-4274-9472-3d5b10ba4358"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.118565 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e2cb3e85-dce2-4274-9472-3d5b10ba4358" (UID: "e2cb3e85-dce2-4274-9472-3d5b10ba4358"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.182417 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.182472 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2cb3e85-dce2-4274-9472-3d5b10ba4358-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.182486 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l792b\" (UniqueName: \"kubernetes.io/projected/e2cb3e85-dce2-4274-9472-3d5b10ba4358-kube-api-access-l792b\") on node \"crc\" DevicePath \"\"" Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.517448 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" event={"ID":"e2cb3e85-dce2-4274-9472-3d5b10ba4358","Type":"ContainerDied","Data":"71bf64fd3bf88e0e027ef821dd0880b12bd13afbfa907f673669d7aa624e5a4c"} Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.517494 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71bf64fd3bf88e0e027ef821dd0880b12bd13afbfa907f673669d7aa624e5a4c" Sep 30 14:55:02 crc kubenswrapper[4799]: I0930 14:55:02.517535 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llf6t" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.033184 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph"] Sep 30 14:55:09 crc kubenswrapper[4799]: E0930 14:55:09.035470 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="extract-content" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.035565 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="extract-content" Sep 30 14:55:09 crc kubenswrapper[4799]: E0930 14:55:09.035652 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerName="registry-server" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.035752 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerName="registry-server" Sep 30 14:55:09 crc kubenswrapper[4799]: E0930 14:55:09.035834 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2cb3e85-dce2-4274-9472-3d5b10ba4358" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.035902 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2cb3e85-dce2-4274-9472-3d5b10ba4358" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:55:09 crc kubenswrapper[4799]: E0930 14:55:09.035979 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerName="extract-utilities" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.036043 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerName="extract-utilities" Sep 30 14:55:09 crc kubenswrapper[4799]: E0930 14:55:09.036121 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerName="extract-content" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.036189 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerName="extract-content" Sep 30 14:55:09 crc kubenswrapper[4799]: E0930 14:55:09.036279 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="registry-server" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.036350 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="registry-server" Sep 30 14:55:09 crc kubenswrapper[4799]: E0930 14:55:09.036431 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="extract-utilities" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.036503 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="extract-utilities" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.036809 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbb077e-8e14-41fd-bc6c-3c47751ded5a" containerName="registry-server" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.036907 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2cb3e85-dce2-4274-9472-3d5b10ba4358" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.036999 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="8277152e-d298-44b1-8dc8-127b6a3349f2" containerName="registry-server" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.037984 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.040924 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.041056 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.041459 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.041690 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.053316 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph"] Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.211979 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2tv4\" (UniqueName: \"kubernetes.io/projected/200d7385-c387-4fc1-b108-fd1873d6da2d-kube-api-access-g2tv4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.212050 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.212268 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.314628 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2tv4\" (UniqueName: \"kubernetes.io/projected/200d7385-c387-4fc1-b108-fd1873d6da2d-kube-api-access-g2tv4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.314718 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.314824 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.323321 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.327291 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.334412 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2tv4\" (UniqueName: \"kubernetes.io/projected/200d7385-c387-4fc1-b108-fd1873d6da2d-kube-api-access-g2tv4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-q5dph\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.361844 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.962683 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph"] Sep 30 14:55:09 crc kubenswrapper[4799]: W0930 14:55:09.966698 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod200d7385_c387_4fc1_b108_fd1873d6da2d.slice/crio-5e39add5e1c8afb9543e061d29ac656b0abb1ba13c495657fe84ed25b8c60cc7 WatchSource:0}: Error finding container 5e39add5e1c8afb9543e061d29ac656b0abb1ba13c495657fe84ed25b8c60cc7: Status 404 returned error can't find the container with id 5e39add5e1c8afb9543e061d29ac656b0abb1ba13c495657fe84ed25b8c60cc7 Sep 30 14:55:09 crc kubenswrapper[4799]: I0930 14:55:09.970138 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 14:55:10 crc kubenswrapper[4799]: I0930 14:55:10.608078 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" event={"ID":"200d7385-c387-4fc1-b108-fd1873d6da2d","Type":"ContainerStarted","Data":"b9ef49a65c2875bea429dfbc5284b6002b37bea020e3e03e8e94f4d48792bd57"} Sep 30 14:55:10 crc kubenswrapper[4799]: I0930 14:55:10.609560 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" event={"ID":"200d7385-c387-4fc1-b108-fd1873d6da2d","Type":"ContainerStarted","Data":"5e39add5e1c8afb9543e061d29ac656b0abb1ba13c495657fe84ed25b8c60cc7"} Sep 30 14:55:10 crc kubenswrapper[4799]: I0930 14:55:10.632225 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" podStartSLOduration=1.43582176 podStartE2EDuration="1.6321949s" podCreationTimestamp="2025-09-30 14:55:09 +0000 UTC" firstStartedPulling="2025-09-30 14:55:09.969840166 +0000 UTC m=+2132.053440593" lastFinishedPulling="2025-09-30 14:55:10.166213306 +0000 UTC m=+2132.249813733" observedRunningTime="2025-09-30 14:55:10.62739013 +0000 UTC m=+2132.710990577" watchObservedRunningTime="2025-09-30 14:55:10.6321949 +0000 UTC m=+2132.715795327" Sep 30 14:55:59 crc kubenswrapper[4799]: I0930 14:55:59.664831 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:55:59 crc kubenswrapper[4799]: I0930 14:55:59.665454 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:56:01 crc kubenswrapper[4799]: I0930 14:56:01.110926 4799 generic.go:334] "Generic (PLEG): container finished" podID="200d7385-c387-4fc1-b108-fd1873d6da2d" containerID="b9ef49a65c2875bea429dfbc5284b6002b37bea020e3e03e8e94f4d48792bd57" exitCode=0 Sep 30 14:56:01 crc kubenswrapper[4799]: I0930 14:56:01.111010 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" event={"ID":"200d7385-c387-4fc1-b108-fd1873d6da2d","Type":"ContainerDied","Data":"b9ef49a65c2875bea429dfbc5284b6002b37bea020e3e03e8e94f4d48792bd57"} Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.555982 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.623155 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-inventory\") pod \"200d7385-c387-4fc1-b108-fd1873d6da2d\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.623208 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2tv4\" (UniqueName: \"kubernetes.io/projected/200d7385-c387-4fc1-b108-fd1873d6da2d-kube-api-access-g2tv4\") pod \"200d7385-c387-4fc1-b108-fd1873d6da2d\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.623278 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-ssh-key\") pod \"200d7385-c387-4fc1-b108-fd1873d6da2d\" (UID: \"200d7385-c387-4fc1-b108-fd1873d6da2d\") " Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.631117 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/200d7385-c387-4fc1-b108-fd1873d6da2d-kube-api-access-g2tv4" (OuterVolumeSpecName: "kube-api-access-g2tv4") pod "200d7385-c387-4fc1-b108-fd1873d6da2d" (UID: "200d7385-c387-4fc1-b108-fd1873d6da2d"). InnerVolumeSpecName "kube-api-access-g2tv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.656389 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "200d7385-c387-4fc1-b108-fd1873d6da2d" (UID: "200d7385-c387-4fc1-b108-fd1873d6da2d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.658927 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-inventory" (OuterVolumeSpecName: "inventory") pod "200d7385-c387-4fc1-b108-fd1873d6da2d" (UID: "200d7385-c387-4fc1-b108-fd1873d6da2d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.725226 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.725265 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2tv4\" (UniqueName: \"kubernetes.io/projected/200d7385-c387-4fc1-b108-fd1873d6da2d-kube-api-access-g2tv4\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:02 crc kubenswrapper[4799]: I0930 14:56:02.725277 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/200d7385-c387-4fc1-b108-fd1873d6da2d-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.133138 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" event={"ID":"200d7385-c387-4fc1-b108-fd1873d6da2d","Type":"ContainerDied","Data":"5e39add5e1c8afb9543e061d29ac656b0abb1ba13c495657fe84ed25b8c60cc7"} Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.133197 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e39add5e1c8afb9543e061d29ac656b0abb1ba13c495657fe84ed25b8c60cc7" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.133242 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-q5dph" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.242884 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8swsg"] Sep 30 14:56:03 crc kubenswrapper[4799]: E0930 14:56:03.243380 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="200d7385-c387-4fc1-b108-fd1873d6da2d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.243405 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="200d7385-c387-4fc1-b108-fd1873d6da2d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.244061 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="200d7385-c387-4fc1-b108-fd1873d6da2d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.244941 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.249821 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.250088 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.250145 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.250154 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.284086 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8swsg"] Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.337518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.338753 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.338852 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6msj\" (UniqueName: \"kubernetes.io/projected/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-kube-api-access-x6msj\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.441183 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.441245 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6msj\" (UniqueName: \"kubernetes.io/projected/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-kube-api-access-x6msj\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.441389 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.447200 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.451389 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.469583 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6msj\" (UniqueName: \"kubernetes.io/projected/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-kube-api-access-x6msj\") pod \"ssh-known-hosts-edpm-deployment-8swsg\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:03 crc kubenswrapper[4799]: I0930 14:56:03.562859 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:04 crc kubenswrapper[4799]: I0930 14:56:04.141487 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8swsg"] Sep 30 14:56:05 crc kubenswrapper[4799]: I0930 14:56:05.155109 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" event={"ID":"5f994a1a-5c13-4a9e-9d44-ea7a8389e943","Type":"ContainerStarted","Data":"2cfcf5de0fb04939d8abbcdc87149de772933470add11510d761a3e629a886b3"} Sep 30 14:56:05 crc kubenswrapper[4799]: I0930 14:56:05.155520 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" event={"ID":"5f994a1a-5c13-4a9e-9d44-ea7a8389e943","Type":"ContainerStarted","Data":"30ed66a2eb12213e18da9399404d9d1a215c4623ffd6c62d7be521ac212529ab"} Sep 30 14:56:05 crc kubenswrapper[4799]: I0930 14:56:05.180213 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" podStartSLOduration=1.9979982870000002 podStartE2EDuration="2.180188094s" podCreationTimestamp="2025-09-30 14:56:03 +0000 UTC" firstStartedPulling="2025-09-30 14:56:04.140777897 +0000 UTC m=+2186.224378324" lastFinishedPulling="2025-09-30 14:56:04.322967704 +0000 UTC m=+2186.406568131" observedRunningTime="2025-09-30 14:56:05.176717083 +0000 UTC m=+2187.260317530" watchObservedRunningTime="2025-09-30 14:56:05.180188094 +0000 UTC m=+2187.263788521" Sep 30 14:56:13 crc kubenswrapper[4799]: I0930 14:56:13.232885 4799 generic.go:334] "Generic (PLEG): container finished" podID="5f994a1a-5c13-4a9e-9d44-ea7a8389e943" containerID="2cfcf5de0fb04939d8abbcdc87149de772933470add11510d761a3e629a886b3" exitCode=0 Sep 30 14:56:13 crc kubenswrapper[4799]: I0930 14:56:13.232972 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" event={"ID":"5f994a1a-5c13-4a9e-9d44-ea7a8389e943","Type":"ContainerDied","Data":"2cfcf5de0fb04939d8abbcdc87149de772933470add11510d761a3e629a886b3"} Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.716432 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.788127 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-inventory-0\") pod \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.789560 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6msj\" (UniqueName: \"kubernetes.io/projected/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-kube-api-access-x6msj\") pod \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.789613 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-ssh-key-openstack-edpm-ipam\") pod \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\" (UID: \"5f994a1a-5c13-4a9e-9d44-ea7a8389e943\") " Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.794388 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-kube-api-access-x6msj" (OuterVolumeSpecName: "kube-api-access-x6msj") pod "5f994a1a-5c13-4a9e-9d44-ea7a8389e943" (UID: "5f994a1a-5c13-4a9e-9d44-ea7a8389e943"). InnerVolumeSpecName "kube-api-access-x6msj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.819427 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5f994a1a-5c13-4a9e-9d44-ea7a8389e943" (UID: "5f994a1a-5c13-4a9e-9d44-ea7a8389e943"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.821356 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5f994a1a-5c13-4a9e-9d44-ea7a8389e943" (UID: "5f994a1a-5c13-4a9e-9d44-ea7a8389e943"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.892851 4799 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-inventory-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.892904 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6msj\" (UniqueName: \"kubernetes.io/projected/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-kube-api-access-x6msj\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:14 crc kubenswrapper[4799]: I0930 14:56:14.892925 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f994a1a-5c13-4a9e-9d44-ea7a8389e943-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.253496 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" event={"ID":"5f994a1a-5c13-4a9e-9d44-ea7a8389e943","Type":"ContainerDied","Data":"30ed66a2eb12213e18da9399404d9d1a215c4623ffd6c62d7be521ac212529ab"} Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.253543 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30ed66a2eb12213e18da9399404d9d1a215c4623ffd6c62d7be521ac212529ab" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.253608 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8swsg" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.347821 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p"] Sep 30 14:56:15 crc kubenswrapper[4799]: E0930 14:56:15.348343 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f994a1a-5c13-4a9e-9d44-ea7a8389e943" containerName="ssh-known-hosts-edpm-deployment" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.348369 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f994a1a-5c13-4a9e-9d44-ea7a8389e943" containerName="ssh-known-hosts-edpm-deployment" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.348594 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f994a1a-5c13-4a9e-9d44-ea7a8389e943" containerName="ssh-known-hosts-edpm-deployment" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.349334 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.352542 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.352849 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.356779 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.357855 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.362172 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p"] Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.502546 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.502626 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgqbs\" (UniqueName: \"kubernetes.io/projected/34ff6ecd-36de-4642-953a-c5d60e8ab31e-kube-api-access-qgqbs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.502699 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.621905 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.621992 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgqbs\" (UniqueName: \"kubernetes.io/projected/34ff6ecd-36de-4642-953a-c5d60e8ab31e-kube-api-access-qgqbs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.622074 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.626883 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.627816 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.642207 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgqbs\" (UniqueName: \"kubernetes.io/projected/34ff6ecd-36de-4642-953a-c5d60e8ab31e-kube-api-access-qgqbs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g782p\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:15 crc kubenswrapper[4799]: I0930 14:56:15.670917 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:16 crc kubenswrapper[4799]: I0930 14:56:16.231080 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p"] Sep 30 14:56:16 crc kubenswrapper[4799]: I0930 14:56:16.264338 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" event={"ID":"34ff6ecd-36de-4642-953a-c5d60e8ab31e","Type":"ContainerStarted","Data":"9ee5ee1f2284a98da58a3a94e27950fcdc0efd34ccafab2ddd5d5298d5424c72"} Sep 30 14:56:17 crc kubenswrapper[4799]: I0930 14:56:17.275203 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" event={"ID":"34ff6ecd-36de-4642-953a-c5d60e8ab31e","Type":"ContainerStarted","Data":"e010d8d73e01d40204e2288e97512ecd1b0f87a88b8b0927be972eaa9d462a54"} Sep 30 14:56:17 crc kubenswrapper[4799]: I0930 14:56:17.298344 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" podStartSLOduration=2.130455105 podStartE2EDuration="2.298300124s" podCreationTimestamp="2025-09-30 14:56:15 +0000 UTC" firstStartedPulling="2025-09-30 14:56:16.240632784 +0000 UTC m=+2198.324233211" lastFinishedPulling="2025-09-30 14:56:16.408477803 +0000 UTC m=+2198.492078230" observedRunningTime="2025-09-30 14:56:17.292009581 +0000 UTC m=+2199.375610008" watchObservedRunningTime="2025-09-30 14:56:17.298300124 +0000 UTC m=+2199.381900551" Sep 30 14:56:26 crc kubenswrapper[4799]: I0930 14:56:26.364935 4799 generic.go:334] "Generic (PLEG): container finished" podID="34ff6ecd-36de-4642-953a-c5d60e8ab31e" containerID="e010d8d73e01d40204e2288e97512ecd1b0f87a88b8b0927be972eaa9d462a54" exitCode=0 Sep 30 14:56:26 crc kubenswrapper[4799]: I0930 14:56:26.365015 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" event={"ID":"34ff6ecd-36de-4642-953a-c5d60e8ab31e","Type":"ContainerDied","Data":"e010d8d73e01d40204e2288e97512ecd1b0f87a88b8b0927be972eaa9d462a54"} Sep 30 14:56:27 crc kubenswrapper[4799]: I0930 14:56:27.870503 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.004016 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-inventory\") pod \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.004155 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-ssh-key\") pod \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.004271 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgqbs\" (UniqueName: \"kubernetes.io/projected/34ff6ecd-36de-4642-953a-c5d60e8ab31e-kube-api-access-qgqbs\") pod \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\" (UID: \"34ff6ecd-36de-4642-953a-c5d60e8ab31e\") " Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.020858 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ff6ecd-36de-4642-953a-c5d60e8ab31e-kube-api-access-qgqbs" (OuterVolumeSpecName: "kube-api-access-qgqbs") pod "34ff6ecd-36de-4642-953a-c5d60e8ab31e" (UID: "34ff6ecd-36de-4642-953a-c5d60e8ab31e"). InnerVolumeSpecName "kube-api-access-qgqbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.041796 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "34ff6ecd-36de-4642-953a-c5d60e8ab31e" (UID: "34ff6ecd-36de-4642-953a-c5d60e8ab31e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.041849 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-inventory" (OuterVolumeSpecName: "inventory") pod "34ff6ecd-36de-4642-953a-c5d60e8ab31e" (UID: "34ff6ecd-36de-4642-953a-c5d60e8ab31e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.106593 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.106637 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgqbs\" (UniqueName: \"kubernetes.io/projected/34ff6ecd-36de-4642-953a-c5d60e8ab31e-kube-api-access-qgqbs\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.106669 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ff6ecd-36de-4642-953a-c5d60e8ab31e-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.389271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" event={"ID":"34ff6ecd-36de-4642-953a-c5d60e8ab31e","Type":"ContainerDied","Data":"9ee5ee1f2284a98da58a3a94e27950fcdc0efd34ccafab2ddd5d5298d5424c72"} Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.389327 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ee5ee1f2284a98da58a3a94e27950fcdc0efd34ccafab2ddd5d5298d5424c72" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.389412 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g782p" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.522957 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n"] Sep 30 14:56:28 crc kubenswrapper[4799]: E0930 14:56:28.523470 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ff6ecd-36de-4642-953a-c5d60e8ab31e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.523491 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ff6ecd-36de-4642-953a-c5d60e8ab31e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.523712 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ff6ecd-36de-4642-953a-c5d60e8ab31e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.524537 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.532527 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n"] Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.535276 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.535478 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.535588 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.538803 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.616145 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.616203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.616510 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g2fn\" (UniqueName: \"kubernetes.io/projected/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-kube-api-access-9g2fn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.718733 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g2fn\" (UniqueName: \"kubernetes.io/projected/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-kube-api-access-9g2fn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.718909 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.718941 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.727712 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.728219 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.747209 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g2fn\" (UniqueName: \"kubernetes.io/projected/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-kube-api-access-9g2fn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:28 crc kubenswrapper[4799]: I0930 14:56:28.848147 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:29 crc kubenswrapper[4799]: I0930 14:56:29.597492 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n"] Sep 30 14:56:29 crc kubenswrapper[4799]: I0930 14:56:29.648837 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:56:29 crc kubenswrapper[4799]: I0930 14:56:29.649213 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:56:30 crc kubenswrapper[4799]: I0930 14:56:30.417218 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" event={"ID":"aadd9e81-92aa-46fe-bae7-8a0ccddd873f","Type":"ContainerStarted","Data":"6abb5459d179b75e62533bef89653349df1d7401a4b3f4dc123bccc1fefe2100"} Sep 30 14:56:30 crc kubenswrapper[4799]: I0930 14:56:30.417993 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" event={"ID":"aadd9e81-92aa-46fe-bae7-8a0ccddd873f","Type":"ContainerStarted","Data":"3d1bf321223d5f3a55b9c81c5e199c092fbaa48cd4e49f19f48b721699ee9fd6"} Sep 30 14:56:30 crc kubenswrapper[4799]: I0930 14:56:30.443482 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" podStartSLOduration=2.215486884 podStartE2EDuration="2.443449104s" podCreationTimestamp="2025-09-30 14:56:28 +0000 UTC" firstStartedPulling="2025-09-30 14:56:29.636110587 +0000 UTC m=+2211.719711014" lastFinishedPulling="2025-09-30 14:56:29.864072807 +0000 UTC m=+2211.947673234" observedRunningTime="2025-09-30 14:56:30.43612222 +0000 UTC m=+2212.519722647" watchObservedRunningTime="2025-09-30 14:56:30.443449104 +0000 UTC m=+2212.527049531" Sep 30 14:56:41 crc kubenswrapper[4799]: I0930 14:56:41.526103 4799 generic.go:334] "Generic (PLEG): container finished" podID="aadd9e81-92aa-46fe-bae7-8a0ccddd873f" containerID="6abb5459d179b75e62533bef89653349df1d7401a4b3f4dc123bccc1fefe2100" exitCode=0 Sep 30 14:56:41 crc kubenswrapper[4799]: I0930 14:56:41.526179 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" event={"ID":"aadd9e81-92aa-46fe-bae7-8a0ccddd873f","Type":"ContainerDied","Data":"6abb5459d179b75e62533bef89653349df1d7401a4b3f4dc123bccc1fefe2100"} Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.013473 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.151088 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-ssh-key\") pod \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.151230 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-inventory\") pod \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.151339 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g2fn\" (UniqueName: \"kubernetes.io/projected/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-kube-api-access-9g2fn\") pod \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\" (UID: \"aadd9e81-92aa-46fe-bae7-8a0ccddd873f\") " Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.158721 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-kube-api-access-9g2fn" (OuterVolumeSpecName: "kube-api-access-9g2fn") pod "aadd9e81-92aa-46fe-bae7-8a0ccddd873f" (UID: "aadd9e81-92aa-46fe-bae7-8a0ccddd873f"). InnerVolumeSpecName "kube-api-access-9g2fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.181751 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-inventory" (OuterVolumeSpecName: "inventory") pod "aadd9e81-92aa-46fe-bae7-8a0ccddd873f" (UID: "aadd9e81-92aa-46fe-bae7-8a0ccddd873f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.191532 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aadd9e81-92aa-46fe-bae7-8a0ccddd873f" (UID: "aadd9e81-92aa-46fe-bae7-8a0ccddd873f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.253842 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.253881 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.253892 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g2fn\" (UniqueName: \"kubernetes.io/projected/aadd9e81-92aa-46fe-bae7-8a0ccddd873f-kube-api-access-9g2fn\") on node \"crc\" DevicePath \"\"" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.553914 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" event={"ID":"aadd9e81-92aa-46fe-bae7-8a0ccddd873f","Type":"ContainerDied","Data":"3d1bf321223d5f3a55b9c81c5e199c092fbaa48cd4e49f19f48b721699ee9fd6"} Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.554149 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d1bf321223d5f3a55b9c81c5e199c092fbaa48cd4e49f19f48b721699ee9fd6" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.554297 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.654321 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb"] Sep 30 14:56:43 crc kubenswrapper[4799]: E0930 14:56:43.654871 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aadd9e81-92aa-46fe-bae7-8a0ccddd873f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.654894 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="aadd9e81-92aa-46fe-bae7-8a0ccddd873f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.655072 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="aadd9e81-92aa-46fe-bae7-8a0ccddd873f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.655946 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.662178 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.662346 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.662609 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.662748 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.663178 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.663288 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.664590 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb"] Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.666727 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.666935 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.767169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.767614 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.767751 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.767859 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.767972 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.768078 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.768231 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.768367 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch5td\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-kube-api-access-ch5td\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.768454 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.768545 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.768765 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.768916 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.769044 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.769177 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.871178 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch5td\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-kube-api-access-ch5td\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.871538 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.871637 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.871819 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.871908 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.871994 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.872081 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.872195 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.872292 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.872535 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.872632 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.873210 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.873334 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.873530 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.877168 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.878250 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.878366 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.879017 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.879289 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.879912 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.880182 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.885509 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.886604 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.887407 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.887754 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.888131 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.891757 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.896595 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch5td\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-kube-api-access-ch5td\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:43 crc kubenswrapper[4799]: I0930 14:56:43.981384 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:56:44 crc kubenswrapper[4799]: I0930 14:56:44.616199 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb"] Sep 30 14:56:45 crc kubenswrapper[4799]: I0930 14:56:45.574998 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" event={"ID":"b410520b-9fd5-436d-9f2d-48a9ae6e517a","Type":"ContainerStarted","Data":"2c82a8c440e8d5feacd25e26c9492fa4839d140c87184701a6990b0a81f84279"} Sep 30 14:56:45 crc kubenswrapper[4799]: I0930 14:56:45.575329 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" event={"ID":"b410520b-9fd5-436d-9f2d-48a9ae6e517a","Type":"ContainerStarted","Data":"cc103956d2c83ab7812a8e07e8c73f0a36f6d518db463552c3c4bfde453445c7"} Sep 30 14:56:45 crc kubenswrapper[4799]: I0930 14:56:45.606855 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" podStartSLOduration=2.41759762 podStartE2EDuration="2.606829203s" podCreationTimestamp="2025-09-30 14:56:43 +0000 UTC" firstStartedPulling="2025-09-30 14:56:44.626896107 +0000 UTC m=+2226.710496534" lastFinishedPulling="2025-09-30 14:56:44.8161277 +0000 UTC m=+2226.899728117" observedRunningTime="2025-09-30 14:56:45.59985983 +0000 UTC m=+2227.683460257" watchObservedRunningTime="2025-09-30 14:56:45.606829203 +0000 UTC m=+2227.690429630" Sep 30 14:56:55 crc kubenswrapper[4799]: I0930 14:56:55.784635 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h9lr8"] Sep 30 14:56:55 crc kubenswrapper[4799]: I0930 14:56:55.788984 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:55 crc kubenswrapper[4799]: I0930 14:56:55.802988 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h9lr8"] Sep 30 14:56:55 crc kubenswrapper[4799]: I0930 14:56:55.933097 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mwzq\" (UniqueName: \"kubernetes.io/projected/352dbbed-8a3c-4047-aae2-6d1953971cc9-kube-api-access-6mwzq\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:55 crc kubenswrapper[4799]: I0930 14:56:55.933258 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-utilities\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:55 crc kubenswrapper[4799]: I0930 14:56:55.933348 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-catalog-content\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:56 crc kubenswrapper[4799]: I0930 14:56:56.035929 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mwzq\" (UniqueName: \"kubernetes.io/projected/352dbbed-8a3c-4047-aae2-6d1953971cc9-kube-api-access-6mwzq\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:56 crc kubenswrapper[4799]: I0930 14:56:56.036035 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-utilities\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:56 crc kubenswrapper[4799]: I0930 14:56:56.036092 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-catalog-content\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:56 crc kubenswrapper[4799]: I0930 14:56:56.036919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-catalog-content\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:56 crc kubenswrapper[4799]: I0930 14:56:56.037275 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-utilities\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:56 crc kubenswrapper[4799]: I0930 14:56:56.064318 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mwzq\" (UniqueName: \"kubernetes.io/projected/352dbbed-8a3c-4047-aae2-6d1953971cc9-kube-api-access-6mwzq\") pod \"certified-operators-h9lr8\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:56 crc kubenswrapper[4799]: I0930 14:56:56.117530 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:56:56 crc kubenswrapper[4799]: I0930 14:56:56.762613 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h9lr8"] Sep 30 14:56:57 crc kubenswrapper[4799]: I0930 14:56:57.725057 4799 generic.go:334] "Generic (PLEG): container finished" podID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerID="13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00" exitCode=0 Sep 30 14:56:57 crc kubenswrapper[4799]: I0930 14:56:57.725153 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h9lr8" event={"ID":"352dbbed-8a3c-4047-aae2-6d1953971cc9","Type":"ContainerDied","Data":"13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00"} Sep 30 14:56:57 crc kubenswrapper[4799]: I0930 14:56:57.725359 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h9lr8" event={"ID":"352dbbed-8a3c-4047-aae2-6d1953971cc9","Type":"ContainerStarted","Data":"38a1c749b4598afdb7823a0936c6bbc926b827e53ccf5d9e9f2dff37bf38f500"} Sep 30 14:56:58 crc kubenswrapper[4799]: I0930 14:56:58.737711 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h9lr8" event={"ID":"352dbbed-8a3c-4047-aae2-6d1953971cc9","Type":"ContainerStarted","Data":"fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f"} Sep 30 14:56:59 crc kubenswrapper[4799]: I0930 14:56:59.648551 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 14:56:59 crc kubenswrapper[4799]: I0930 14:56:59.649134 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 14:56:59 crc kubenswrapper[4799]: I0930 14:56:59.649217 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 14:56:59 crc kubenswrapper[4799]: I0930 14:56:59.650281 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 14:56:59 crc kubenswrapper[4799]: I0930 14:56:59.650358 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" gracePeriod=600 Sep 30 14:56:59 crc kubenswrapper[4799]: I0930 14:56:59.751742 4799 generic.go:334] "Generic (PLEG): container finished" podID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerID="fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f" exitCode=0 Sep 30 14:56:59 crc kubenswrapper[4799]: I0930 14:56:59.751805 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h9lr8" event={"ID":"352dbbed-8a3c-4047-aae2-6d1953971cc9","Type":"ContainerDied","Data":"fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f"} Sep 30 14:56:59 crc kubenswrapper[4799]: E0930 14:56:59.782368 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:57:00 crc kubenswrapper[4799]: I0930 14:57:00.767698 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" exitCode=0 Sep 30 14:57:00 crc kubenswrapper[4799]: I0930 14:57:00.767796 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976"} Sep 30 14:57:00 crc kubenswrapper[4799]: I0930 14:57:00.768425 4799 scope.go:117] "RemoveContainer" containerID="116aa297eadbb7887d7d899e95966cb2028cb7fa70dc029dd1713cd3c032c2a4" Sep 30 14:57:00 crc kubenswrapper[4799]: I0930 14:57:00.769475 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:57:00 crc kubenswrapper[4799]: E0930 14:57:00.769921 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:57:00 crc kubenswrapper[4799]: I0930 14:57:00.774677 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h9lr8" event={"ID":"352dbbed-8a3c-4047-aae2-6d1953971cc9","Type":"ContainerStarted","Data":"8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e"} Sep 30 14:57:00 crc kubenswrapper[4799]: I0930 14:57:00.830059 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h9lr8" podStartSLOduration=3.158255726 podStartE2EDuration="5.830033046s" podCreationTimestamp="2025-09-30 14:56:55 +0000 UTC" firstStartedPulling="2025-09-30 14:56:57.727293031 +0000 UTC m=+2239.810893468" lastFinishedPulling="2025-09-30 14:57:00.399070361 +0000 UTC m=+2242.482670788" observedRunningTime="2025-09-30 14:57:00.829177291 +0000 UTC m=+2242.912777718" watchObservedRunningTime="2025-09-30 14:57:00.830033046 +0000 UTC m=+2242.913633493" Sep 30 14:57:06 crc kubenswrapper[4799]: I0930 14:57:06.117880 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:57:06 crc kubenswrapper[4799]: I0930 14:57:06.118593 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:57:06 crc kubenswrapper[4799]: I0930 14:57:06.177573 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:57:06 crc kubenswrapper[4799]: I0930 14:57:06.880345 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:57:06 crc kubenswrapper[4799]: I0930 14:57:06.940834 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h9lr8"] Sep 30 14:57:08 crc kubenswrapper[4799]: I0930 14:57:08.849826 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h9lr8" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerName="registry-server" containerID="cri-o://8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e" gracePeriod=2 Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.377661 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.544807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mwzq\" (UniqueName: \"kubernetes.io/projected/352dbbed-8a3c-4047-aae2-6d1953971cc9-kube-api-access-6mwzq\") pod \"352dbbed-8a3c-4047-aae2-6d1953971cc9\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.545019 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-catalog-content\") pod \"352dbbed-8a3c-4047-aae2-6d1953971cc9\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.545042 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-utilities\") pod \"352dbbed-8a3c-4047-aae2-6d1953971cc9\" (UID: \"352dbbed-8a3c-4047-aae2-6d1953971cc9\") " Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.546148 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-utilities" (OuterVolumeSpecName: "utilities") pod "352dbbed-8a3c-4047-aae2-6d1953971cc9" (UID: "352dbbed-8a3c-4047-aae2-6d1953971cc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.574787 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352dbbed-8a3c-4047-aae2-6d1953971cc9-kube-api-access-6mwzq" (OuterVolumeSpecName: "kube-api-access-6mwzq") pod "352dbbed-8a3c-4047-aae2-6d1953971cc9" (UID: "352dbbed-8a3c-4047-aae2-6d1953971cc9"). InnerVolumeSpecName "kube-api-access-6mwzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.647690 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.647745 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mwzq\" (UniqueName: \"kubernetes.io/projected/352dbbed-8a3c-4047-aae2-6d1953971cc9-kube-api-access-6mwzq\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.863098 4799 generic.go:334] "Generic (PLEG): container finished" podID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerID="8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e" exitCode=0 Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.863181 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h9lr8" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.863235 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h9lr8" event={"ID":"352dbbed-8a3c-4047-aae2-6d1953971cc9","Type":"ContainerDied","Data":"8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e"} Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.864825 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h9lr8" event={"ID":"352dbbed-8a3c-4047-aae2-6d1953971cc9","Type":"ContainerDied","Data":"38a1c749b4598afdb7823a0936c6bbc926b827e53ccf5d9e9f2dff37bf38f500"} Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.864891 4799 scope.go:117] "RemoveContainer" containerID="8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.889568 4799 scope.go:117] "RemoveContainer" containerID="fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.937798 4799 scope.go:117] "RemoveContainer" containerID="13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.975116 4799 scope.go:117] "RemoveContainer" containerID="8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e" Sep 30 14:57:09 crc kubenswrapper[4799]: E0930 14:57:09.977721 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e\": container with ID starting with 8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e not found: ID does not exist" containerID="8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.977772 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e"} err="failed to get container status \"8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e\": rpc error: code = NotFound desc = could not find container \"8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e\": container with ID starting with 8ed38e69850c4bc551982212477cebcdc71d608cc1125eaf1b9811ea53062b7e not found: ID does not exist" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.977807 4799 scope.go:117] "RemoveContainer" containerID="fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f" Sep 30 14:57:09 crc kubenswrapper[4799]: E0930 14:57:09.978973 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f\": container with ID starting with fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f not found: ID does not exist" containerID="fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.979093 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f"} err="failed to get container status \"fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f\": rpc error: code = NotFound desc = could not find container \"fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f\": container with ID starting with fc92b79ba9717e1ab463aa27af894fd602f61e1480df2447685a70a4a8584e6f not found: ID does not exist" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.979128 4799 scope.go:117] "RemoveContainer" containerID="13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00" Sep 30 14:57:09 crc kubenswrapper[4799]: E0930 14:57:09.979730 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00\": container with ID starting with 13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00 not found: ID does not exist" containerID="13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00" Sep 30 14:57:09 crc kubenswrapper[4799]: I0930 14:57:09.979990 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00"} err="failed to get container status \"13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00\": rpc error: code = NotFound desc = could not find container \"13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00\": container with ID starting with 13c4aa6757c82b3b5f1bfe4a46080de630ea7a8f4e52b44f4b422febee716f00 not found: ID does not exist" Sep 30 14:57:10 crc kubenswrapper[4799]: I0930 14:57:10.390851 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "352dbbed-8a3c-4047-aae2-6d1953971cc9" (UID: "352dbbed-8a3c-4047-aae2-6d1953971cc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 14:57:10 crc kubenswrapper[4799]: I0930 14:57:10.469984 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/352dbbed-8a3c-4047-aae2-6d1953971cc9-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:10 crc kubenswrapper[4799]: I0930 14:57:10.515806 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h9lr8"] Sep 30 14:57:10 crc kubenswrapper[4799]: I0930 14:57:10.520539 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h9lr8"] Sep 30 14:57:10 crc kubenswrapper[4799]: E0930 14:57:10.742527 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod352dbbed_8a3c_4047_aae2_6d1953971cc9.slice/crio-38a1c749b4598afdb7823a0936c6bbc926b827e53ccf5d9e9f2dff37bf38f500\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod352dbbed_8a3c_4047_aae2_6d1953971cc9.slice\": RecentStats: unable to find data in memory cache]" Sep 30 14:57:12 crc kubenswrapper[4799]: I0930 14:57:12.519882 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" path="/var/lib/kubelet/pods/352dbbed-8a3c-4047-aae2-6d1953971cc9/volumes" Sep 30 14:57:16 crc kubenswrapper[4799]: I0930 14:57:16.504880 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:57:16 crc kubenswrapper[4799]: E0930 14:57:16.505724 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:57:28 crc kubenswrapper[4799]: I0930 14:57:28.054604 4799 generic.go:334] "Generic (PLEG): container finished" podID="b410520b-9fd5-436d-9f2d-48a9ae6e517a" containerID="2c82a8c440e8d5feacd25e26c9492fa4839d140c87184701a6990b0a81f84279" exitCode=0 Sep 30 14:57:28 crc kubenswrapper[4799]: I0930 14:57:28.054715 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" event={"ID":"b410520b-9fd5-436d-9f2d-48a9ae6e517a","Type":"ContainerDied","Data":"2c82a8c440e8d5feacd25e26c9492fa4839d140c87184701a6990b0a81f84279"} Sep 30 14:57:28 crc kubenswrapper[4799]: I0930 14:57:28.512508 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:57:28 crc kubenswrapper[4799]: E0930 14:57:28.512829 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.537336 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.669121 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-libvirt-combined-ca-bundle\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.669605 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.669770 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.669951 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.670076 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch5td\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-kube-api-access-ch5td\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.670190 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-bootstrap-combined-ca-bundle\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.670297 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-repo-setup-combined-ca-bundle\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.670412 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-neutron-metadata-combined-ca-bundle\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.670525 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-telemetry-combined-ca-bundle\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.670638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ovn-combined-ca-bundle\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.671186 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-inventory\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.671285 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.671381 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-nova-combined-ca-bundle\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.671496 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ssh-key\") pod \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\" (UID: \"b410520b-9fd5-436d-9f2d-48a9ae6e517a\") " Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.696437 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.696509 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.697133 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-kube-api-access-ch5td" (OuterVolumeSpecName: "kube-api-access-ch5td") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "kube-api-access-ch5td". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.697241 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.702882 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.703298 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.703382 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.703681 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.704818 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.707962 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.708107 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.717106 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.735581 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-inventory" (OuterVolumeSpecName: "inventory") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.738504 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b410520b-9fd5-436d-9f2d-48a9ae6e517a" (UID: "b410520b-9fd5-436d-9f2d-48a9ae6e517a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774233 4799 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774468 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774564 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774627 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774711 4799 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774769 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774834 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774890 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.774955 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.775038 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.775100 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch5td\" (UniqueName: \"kubernetes.io/projected/b410520b-9fd5-436d-9f2d-48a9ae6e517a-kube-api-access-ch5td\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.775164 4799 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.775250 4799 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:29 crc kubenswrapper[4799]: I0930 14:57:29.775315 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b410520b-9fd5-436d-9f2d-48a9ae6e517a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.078192 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" event={"ID":"b410520b-9fd5-436d-9f2d-48a9ae6e517a","Type":"ContainerDied","Data":"cc103956d2c83ab7812a8e07e8c73f0a36f6d518db463552c3c4bfde453445c7"} Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.078267 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc103956d2c83ab7812a8e07e8c73f0a36f6d518db463552c3c4bfde453445c7" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.078305 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.211924 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk"] Sep 30 14:57:30 crc kubenswrapper[4799]: E0930 14:57:30.212562 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerName="extract-utilities" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.212586 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerName="extract-utilities" Sep 30 14:57:30 crc kubenswrapper[4799]: E0930 14:57:30.212598 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b410520b-9fd5-436d-9f2d-48a9ae6e517a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.212606 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b410520b-9fd5-436d-9f2d-48a9ae6e517a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 14:57:30 crc kubenswrapper[4799]: E0930 14:57:30.212666 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerName="extract-content" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.212676 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerName="extract-content" Sep 30 14:57:30 crc kubenswrapper[4799]: E0930 14:57:30.212700 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerName="registry-server" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.212706 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerName="registry-server" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.212905 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b410520b-9fd5-436d-9f2d-48a9ae6e517a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.212923 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="352dbbed-8a3c-4047-aae2-6d1953971cc9" containerName="registry-server" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.213805 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.217151 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.221980 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.222904 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.223187 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.223635 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.249127 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk"] Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.285129 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7110fd32-977a-4666-8dc1-d0d217767662-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.285317 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgmvc\" (UniqueName: \"kubernetes.io/projected/7110fd32-977a-4666-8dc1-d0d217767662-kube-api-access-hgmvc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.285360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.285410 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.285500 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.388120 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.388244 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.388387 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.388671 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7110fd32-977a-4666-8dc1-d0d217767662-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.388850 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgmvc\" (UniqueName: \"kubernetes.io/projected/7110fd32-977a-4666-8dc1-d0d217767662-kube-api-access-hgmvc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.391127 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7110fd32-977a-4666-8dc1-d0d217767662-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.393969 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.394803 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.395137 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.408769 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgmvc\" (UniqueName: \"kubernetes.io/projected/7110fd32-977a-4666-8dc1-d0d217767662-kube-api-access-hgmvc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-499wk\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:30 crc kubenswrapper[4799]: I0930 14:57:30.556147 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:57:31 crc kubenswrapper[4799]: I0930 14:57:31.178143 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk"] Sep 30 14:57:32 crc kubenswrapper[4799]: I0930 14:57:32.104327 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" event={"ID":"7110fd32-977a-4666-8dc1-d0d217767662","Type":"ContainerStarted","Data":"25b41c06c4f9c8ec2bb65e580fc6e6d07e78009e1bbcdb26811f5237de53c05c"} Sep 30 14:57:32 crc kubenswrapper[4799]: I0930 14:57:32.104877 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" event={"ID":"7110fd32-977a-4666-8dc1-d0d217767662","Type":"ContainerStarted","Data":"e373001effaa176a0b15ca86eeca2b794c2e06be5530491fd0108b1a1b493a2e"} Sep 30 14:57:32 crc kubenswrapper[4799]: I0930 14:57:32.130548 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" podStartSLOduration=1.9280849899999999 podStartE2EDuration="2.130515056s" podCreationTimestamp="2025-09-30 14:57:30 +0000 UTC" firstStartedPulling="2025-09-30 14:57:31.202890575 +0000 UTC m=+2273.286491002" lastFinishedPulling="2025-09-30 14:57:31.405320641 +0000 UTC m=+2273.488921068" observedRunningTime="2025-09-30 14:57:32.120907516 +0000 UTC m=+2274.204507953" watchObservedRunningTime="2025-09-30 14:57:32.130515056 +0000 UTC m=+2274.214115483" Sep 30 14:57:43 crc kubenswrapper[4799]: I0930 14:57:43.504808 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:57:43 crc kubenswrapper[4799]: E0930 14:57:43.505703 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:57:57 crc kubenswrapper[4799]: I0930 14:57:57.505008 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:57:57 crc kubenswrapper[4799]: E0930 14:57:57.506096 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:58:09 crc kubenswrapper[4799]: I0930 14:58:09.503214 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:58:09 crc kubenswrapper[4799]: E0930 14:58:09.504000 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:58:21 crc kubenswrapper[4799]: I0930 14:58:21.503057 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:58:21 crc kubenswrapper[4799]: E0930 14:58:21.504003 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:58:33 crc kubenswrapper[4799]: I0930 14:58:33.504283 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:58:33 crc kubenswrapper[4799]: E0930 14:58:33.505077 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:58:39 crc kubenswrapper[4799]: I0930 14:58:39.792634 4799 generic.go:334] "Generic (PLEG): container finished" podID="7110fd32-977a-4666-8dc1-d0d217767662" containerID="25b41c06c4f9c8ec2bb65e580fc6e6d07e78009e1bbcdb26811f5237de53c05c" exitCode=0 Sep 30 14:58:39 crc kubenswrapper[4799]: I0930 14:58:39.792745 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" event={"ID":"7110fd32-977a-4666-8dc1-d0d217767662","Type":"ContainerDied","Data":"25b41c06c4f9c8ec2bb65e580fc6e6d07e78009e1bbcdb26811f5237de53c05c"} Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.258342 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.361409 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ovn-combined-ca-bundle\") pod \"7110fd32-977a-4666-8dc1-d0d217767662\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.361533 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ssh-key\") pod \"7110fd32-977a-4666-8dc1-d0d217767662\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.361638 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgmvc\" (UniqueName: \"kubernetes.io/projected/7110fd32-977a-4666-8dc1-d0d217767662-kube-api-access-hgmvc\") pod \"7110fd32-977a-4666-8dc1-d0d217767662\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.361714 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-inventory\") pod \"7110fd32-977a-4666-8dc1-d0d217767662\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.361760 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7110fd32-977a-4666-8dc1-d0d217767662-ovncontroller-config-0\") pod \"7110fd32-977a-4666-8dc1-d0d217767662\" (UID: \"7110fd32-977a-4666-8dc1-d0d217767662\") " Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.367894 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7110fd32-977a-4666-8dc1-d0d217767662-kube-api-access-hgmvc" (OuterVolumeSpecName: "kube-api-access-hgmvc") pod "7110fd32-977a-4666-8dc1-d0d217767662" (UID: "7110fd32-977a-4666-8dc1-d0d217767662"). InnerVolumeSpecName "kube-api-access-hgmvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.367912 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7110fd32-977a-4666-8dc1-d0d217767662" (UID: "7110fd32-977a-4666-8dc1-d0d217767662"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.392053 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-inventory" (OuterVolumeSpecName: "inventory") pod "7110fd32-977a-4666-8dc1-d0d217767662" (UID: "7110fd32-977a-4666-8dc1-d0d217767662"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.392495 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7110fd32-977a-4666-8dc1-d0d217767662-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "7110fd32-977a-4666-8dc1-d0d217767662" (UID: "7110fd32-977a-4666-8dc1-d0d217767662"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.396571 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7110fd32-977a-4666-8dc1-d0d217767662" (UID: "7110fd32-977a-4666-8dc1-d0d217767662"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.464685 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgmvc\" (UniqueName: \"kubernetes.io/projected/7110fd32-977a-4666-8dc1-d0d217767662-kube-api-access-hgmvc\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.464726 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.464739 4799 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7110fd32-977a-4666-8dc1-d0d217767662-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.464755 4799 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.464766 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7110fd32-977a-4666-8dc1-d0d217767662-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.813993 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" event={"ID":"7110fd32-977a-4666-8dc1-d0d217767662","Type":"ContainerDied","Data":"e373001effaa176a0b15ca86eeca2b794c2e06be5530491fd0108b1a1b493a2e"} Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.814045 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e373001effaa176a0b15ca86eeca2b794c2e06be5530491fd0108b1a1b493a2e" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.814087 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-499wk" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.940458 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6"] Sep 30 14:58:41 crc kubenswrapper[4799]: E0930 14:58:41.941272 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7110fd32-977a-4666-8dc1-d0d217767662" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.941302 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7110fd32-977a-4666-8dc1-d0d217767662" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.941559 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7110fd32-977a-4666-8dc1-d0d217767662" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.942398 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.945059 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.946422 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.947145 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.947207 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.947218 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.947825 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:58:41 crc kubenswrapper[4799]: I0930 14:58:41.965712 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6"] Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.079151 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.079384 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.079498 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.079535 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vq6f\" (UniqueName: \"kubernetes.io/projected/4ead1575-9ec6-41d9-8eda-f986d3a28713-kube-api-access-7vq6f\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.079704 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.079755 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.181834 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vq6f\" (UniqueName: \"kubernetes.io/projected/4ead1575-9ec6-41d9-8eda-f986d3a28713-kube-api-access-7vq6f\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.182207 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.182262 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.182316 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.183262 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.183375 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.187829 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.188314 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.188580 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.189897 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.191452 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.206482 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vq6f\" (UniqueName: \"kubernetes.io/projected/4ead1575-9ec6-41d9-8eda-f986d3a28713-kube-api-access-7vq6f\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.262221 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.812833 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6"] Sep 30 14:58:42 crc kubenswrapper[4799]: I0930 14:58:42.825731 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" event={"ID":"4ead1575-9ec6-41d9-8eda-f986d3a28713","Type":"ContainerStarted","Data":"083d545196f750cf80589605b8965dbea07e0ef7430f3beaa12054b5b023f0d5"} Sep 30 14:58:43 crc kubenswrapper[4799]: I0930 14:58:43.836864 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" event={"ID":"4ead1575-9ec6-41d9-8eda-f986d3a28713","Type":"ContainerStarted","Data":"d4c384bc71dd498c0c9bd919195efa3eae9ff3458dabc898eebcc7844075c4b7"} Sep 30 14:58:43 crc kubenswrapper[4799]: I0930 14:58:43.863691 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" podStartSLOduration=2.681489646 podStartE2EDuration="2.863667233s" podCreationTimestamp="2025-09-30 14:58:41 +0000 UTC" firstStartedPulling="2025-09-30 14:58:42.815947922 +0000 UTC m=+2344.899548349" lastFinishedPulling="2025-09-30 14:58:42.998125509 +0000 UTC m=+2345.081725936" observedRunningTime="2025-09-30 14:58:43.855028691 +0000 UTC m=+2345.938629138" watchObservedRunningTime="2025-09-30 14:58:43.863667233 +0000 UTC m=+2345.947267660" Sep 30 14:58:48 crc kubenswrapper[4799]: I0930 14:58:48.511456 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:58:48 crc kubenswrapper[4799]: E0930 14:58:48.512532 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:59:03 crc kubenswrapper[4799]: I0930 14:59:03.503786 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:59:03 crc kubenswrapper[4799]: E0930 14:59:03.504569 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:59:16 crc kubenswrapper[4799]: I0930 14:59:16.504049 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:59:16 crc kubenswrapper[4799]: E0930 14:59:16.505025 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:59:29 crc kubenswrapper[4799]: I0930 14:59:29.504305 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:59:29 crc kubenswrapper[4799]: E0930 14:59:29.505128 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:59:36 crc kubenswrapper[4799]: I0930 14:59:36.336495 4799 generic.go:334] "Generic (PLEG): container finished" podID="4ead1575-9ec6-41d9-8eda-f986d3a28713" containerID="d4c384bc71dd498c0c9bd919195efa3eae9ff3458dabc898eebcc7844075c4b7" exitCode=0 Sep 30 14:59:36 crc kubenswrapper[4799]: I0930 14:59:36.336593 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" event={"ID":"4ead1575-9ec6-41d9-8eda-f986d3a28713","Type":"ContainerDied","Data":"d4c384bc71dd498c0c9bd919195efa3eae9ff3458dabc898eebcc7844075c4b7"} Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.820289 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.848144 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vq6f\" (UniqueName: \"kubernetes.io/projected/4ead1575-9ec6-41d9-8eda-f986d3a28713-kube-api-access-7vq6f\") pod \"4ead1575-9ec6-41d9-8eda-f986d3a28713\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.849031 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4ead1575-9ec6-41d9-8eda-f986d3a28713\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.849183 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-metadata-combined-ca-bundle\") pod \"4ead1575-9ec6-41d9-8eda-f986d3a28713\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.849247 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-nova-metadata-neutron-config-0\") pod \"4ead1575-9ec6-41d9-8eda-f986d3a28713\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.849308 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-ssh-key\") pod \"4ead1575-9ec6-41d9-8eda-f986d3a28713\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.849486 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-inventory\") pod \"4ead1575-9ec6-41d9-8eda-f986d3a28713\" (UID: \"4ead1575-9ec6-41d9-8eda-f986d3a28713\") " Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.856127 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ead1575-9ec6-41d9-8eda-f986d3a28713-kube-api-access-7vq6f" (OuterVolumeSpecName: "kube-api-access-7vq6f") pod "4ead1575-9ec6-41d9-8eda-f986d3a28713" (UID: "4ead1575-9ec6-41d9-8eda-f986d3a28713"). InnerVolumeSpecName "kube-api-access-7vq6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.863219 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4ead1575-9ec6-41d9-8eda-f986d3a28713" (UID: "4ead1575-9ec6-41d9-8eda-f986d3a28713"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.899019 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4ead1575-9ec6-41d9-8eda-f986d3a28713" (UID: "4ead1575-9ec6-41d9-8eda-f986d3a28713"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.910845 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-inventory" (OuterVolumeSpecName: "inventory") pod "4ead1575-9ec6-41d9-8eda-f986d3a28713" (UID: "4ead1575-9ec6-41d9-8eda-f986d3a28713"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.913667 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4ead1575-9ec6-41d9-8eda-f986d3a28713" (UID: "4ead1575-9ec6-41d9-8eda-f986d3a28713"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.914005 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4ead1575-9ec6-41d9-8eda-f986d3a28713" (UID: "4ead1575-9ec6-41d9-8eda-f986d3a28713"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.953501 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.953551 4799 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.953567 4799 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.953580 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.953593 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ead1575-9ec6-41d9-8eda-f986d3a28713-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 14:59:37 crc kubenswrapper[4799]: I0930 14:59:37.953609 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vq6f\" (UniqueName: \"kubernetes.io/projected/4ead1575-9ec6-41d9-8eda-f986d3a28713-kube-api-access-7vq6f\") on node \"crc\" DevicePath \"\"" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.360610 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" event={"ID":"4ead1575-9ec6-41d9-8eda-f986d3a28713","Type":"ContainerDied","Data":"083d545196f750cf80589605b8965dbea07e0ef7430f3beaa12054b5b023f0d5"} Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.360698 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="083d545196f750cf80589605b8965dbea07e0ef7430f3beaa12054b5b023f0d5" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.360739 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.475482 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw"] Sep 30 14:59:38 crc kubenswrapper[4799]: E0930 14:59:38.475945 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ead1575-9ec6-41d9-8eda-f986d3a28713" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.476084 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ead1575-9ec6-41d9-8eda-f986d3a28713" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.476321 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ead1575-9ec6-41d9-8eda-f986d3a28713" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.477028 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.480258 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.480770 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.482578 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.483403 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.494452 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.570872 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.570925 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.571024 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw9vt\" (UniqueName: \"kubernetes.io/projected/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-kube-api-access-mw9vt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.571086 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.571411 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.574974 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw"] Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.672949 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.673008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.673113 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw9vt\" (UniqueName: \"kubernetes.io/projected/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-kube-api-access-mw9vt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.673179 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.673326 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.679960 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.680331 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.680575 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.680706 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.698801 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw9vt\" (UniqueName: \"kubernetes.io/projected/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-kube-api-access-mw9vt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:38 crc kubenswrapper[4799]: I0930 14:59:38.805372 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 14:59:39 crc kubenswrapper[4799]: I0930 14:59:39.409879 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw"] Sep 30 14:59:40 crc kubenswrapper[4799]: I0930 14:59:40.383970 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" event={"ID":"43ae8f90-c778-47e6-8c69-8cdd2b4fa949","Type":"ContainerStarted","Data":"9d069c373284b5539ad2a9103fbee9e88bb6ea9ef45fb434baba875b8fdf2c48"} Sep 30 14:59:40 crc kubenswrapper[4799]: I0930 14:59:40.384514 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" event={"ID":"43ae8f90-c778-47e6-8c69-8cdd2b4fa949","Type":"ContainerStarted","Data":"b980eafbbfeca39d1ab89bf46cb6727a5a74ec91f67ac8b4dd6198f1de057fa4"} Sep 30 14:59:42 crc kubenswrapper[4799]: I0930 14:59:42.506311 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:59:42 crc kubenswrapper[4799]: E0930 14:59:42.506910 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 14:59:56 crc kubenswrapper[4799]: I0930 14:59:56.504416 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 14:59:56 crc kubenswrapper[4799]: E0930 14:59:56.505275 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.145382 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" podStartSLOduration=21.800141065 podStartE2EDuration="22.145334926s" podCreationTimestamp="2025-09-30 14:59:38 +0000 UTC" firstStartedPulling="2025-09-30 14:59:39.415932878 +0000 UTC m=+2401.499533305" lastFinishedPulling="2025-09-30 14:59:39.761126739 +0000 UTC m=+2401.844727166" observedRunningTime="2025-09-30 14:59:40.407196721 +0000 UTC m=+2402.490797148" watchObservedRunningTime="2025-09-30 15:00:00.145334926 +0000 UTC m=+2422.228935353" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.153738 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd"] Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.155117 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.157685 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.158032 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.172017 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd"] Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.261550 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06752d82-47d3-4e68-9db7-b20b7d3bc86d-config-volume\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.261615 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc5mv\" (UniqueName: \"kubernetes.io/projected/06752d82-47d3-4e68-9db7-b20b7d3bc86d-kube-api-access-bc5mv\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.261727 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06752d82-47d3-4e68-9db7-b20b7d3bc86d-secret-volume\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.364294 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06752d82-47d3-4e68-9db7-b20b7d3bc86d-config-volume\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.364685 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc5mv\" (UniqueName: \"kubernetes.io/projected/06752d82-47d3-4e68-9db7-b20b7d3bc86d-kube-api-access-bc5mv\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.364818 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06752d82-47d3-4e68-9db7-b20b7d3bc86d-secret-volume\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.365466 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06752d82-47d3-4e68-9db7-b20b7d3bc86d-config-volume\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.371387 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06752d82-47d3-4e68-9db7-b20b7d3bc86d-secret-volume\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.398999 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc5mv\" (UniqueName: \"kubernetes.io/projected/06752d82-47d3-4e68-9db7-b20b7d3bc86d-kube-api-access-bc5mv\") pod \"collect-profiles-29320740-fksrd\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:00 crc kubenswrapper[4799]: I0930 15:00:00.483061 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:01 crc kubenswrapper[4799]: I0930 15:00:01.208564 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd"] Sep 30 15:00:01 crc kubenswrapper[4799]: I0930 15:00:01.655702 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" event={"ID":"06752d82-47d3-4e68-9db7-b20b7d3bc86d","Type":"ContainerStarted","Data":"6dc59e47406c9988c74d995f8d85982e098d36946dac587fdec43104e35ed64f"} Sep 30 15:00:01 crc kubenswrapper[4799]: I0930 15:00:01.656037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" event={"ID":"06752d82-47d3-4e68-9db7-b20b7d3bc86d","Type":"ContainerStarted","Data":"ea253ee742ab1e11400dcbf3ffc0ce357c724ae8c2bcada337667687b97560ae"} Sep 30 15:00:01 crc kubenswrapper[4799]: I0930 15:00:01.682307 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" podStartSLOduration=1.682275766 podStartE2EDuration="1.682275766s" podCreationTimestamp="2025-09-30 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:00:01.674618841 +0000 UTC m=+2423.758219288" watchObservedRunningTime="2025-09-30 15:00:01.682275766 +0000 UTC m=+2423.765876203" Sep 30 15:00:02 crc kubenswrapper[4799]: I0930 15:00:02.666054 4799 generic.go:334] "Generic (PLEG): container finished" podID="06752d82-47d3-4e68-9db7-b20b7d3bc86d" containerID="6dc59e47406c9988c74d995f8d85982e098d36946dac587fdec43104e35ed64f" exitCode=0 Sep 30 15:00:02 crc kubenswrapper[4799]: I0930 15:00:02.667255 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" event={"ID":"06752d82-47d3-4e68-9db7-b20b7d3bc86d","Type":"ContainerDied","Data":"6dc59e47406c9988c74d995f8d85982e098d36946dac587fdec43104e35ed64f"} Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.059286 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.246439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc5mv\" (UniqueName: \"kubernetes.io/projected/06752d82-47d3-4e68-9db7-b20b7d3bc86d-kube-api-access-bc5mv\") pod \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.246505 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06752d82-47d3-4e68-9db7-b20b7d3bc86d-config-volume\") pod \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.246807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06752d82-47d3-4e68-9db7-b20b7d3bc86d-secret-volume\") pod \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\" (UID: \"06752d82-47d3-4e68-9db7-b20b7d3bc86d\") " Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.247376 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06752d82-47d3-4e68-9db7-b20b7d3bc86d-config-volume" (OuterVolumeSpecName: "config-volume") pod "06752d82-47d3-4e68-9db7-b20b7d3bc86d" (UID: "06752d82-47d3-4e68-9db7-b20b7d3bc86d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.254529 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06752d82-47d3-4e68-9db7-b20b7d3bc86d-kube-api-access-bc5mv" (OuterVolumeSpecName: "kube-api-access-bc5mv") pod "06752d82-47d3-4e68-9db7-b20b7d3bc86d" (UID: "06752d82-47d3-4e68-9db7-b20b7d3bc86d"). InnerVolumeSpecName "kube-api-access-bc5mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.259959 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06752d82-47d3-4e68-9db7-b20b7d3bc86d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "06752d82-47d3-4e68-9db7-b20b7d3bc86d" (UID: "06752d82-47d3-4e68-9db7-b20b7d3bc86d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.302962 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t"] Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.312272 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320695-r4v2t"] Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.349438 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc5mv\" (UniqueName: \"kubernetes.io/projected/06752d82-47d3-4e68-9db7-b20b7d3bc86d-kube-api-access-bc5mv\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.349799 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06752d82-47d3-4e68-9db7-b20b7d3bc86d-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.349883 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06752d82-47d3-4e68-9db7-b20b7d3bc86d-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.516729 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f147ef4-6fba-40b2-bdda-f19618d512eb" path="/var/lib/kubelet/pods/7f147ef4-6fba-40b2-bdda-f19618d512eb/volumes" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.719458 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" event={"ID":"06752d82-47d3-4e68-9db7-b20b7d3bc86d","Type":"ContainerDied","Data":"ea253ee742ab1e11400dcbf3ffc0ce357c724ae8c2bcada337667687b97560ae"} Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.719510 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea253ee742ab1e11400dcbf3ffc0ce357c724ae8c2bcada337667687b97560ae" Sep 30 15:00:04 crc kubenswrapper[4799]: I0930 15:00:04.719608 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd" Sep 30 15:00:11 crc kubenswrapper[4799]: I0930 15:00:11.504032 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:00:11 crc kubenswrapper[4799]: E0930 15:00:11.504902 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:00:12 crc kubenswrapper[4799]: I0930 15:00:12.109738 4799 scope.go:117] "RemoveContainer" containerID="fff6ff50d0661449750a3403f5d699e084f7b63e133787d92d17e96c512110dd" Sep 30 15:00:25 crc kubenswrapper[4799]: I0930 15:00:25.503879 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:00:25 crc kubenswrapper[4799]: E0930 15:00:25.504683 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:00:38 crc kubenswrapper[4799]: I0930 15:00:38.511477 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:00:38 crc kubenswrapper[4799]: E0930 15:00:38.512678 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:00:53 crc kubenswrapper[4799]: I0930 15:00:53.503611 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:00:53 crc kubenswrapper[4799]: E0930 15:00:53.504615 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.159554 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29320741-x298c"] Sep 30 15:01:00 crc kubenswrapper[4799]: E0930 15:01:00.166129 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06752d82-47d3-4e68-9db7-b20b7d3bc86d" containerName="collect-profiles" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.166373 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="06752d82-47d3-4e68-9db7-b20b7d3bc86d" containerName="collect-profiles" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.166779 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="06752d82-47d3-4e68-9db7-b20b7d3bc86d" containerName="collect-profiles" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.168007 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.174828 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29320741-x298c"] Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.257465 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-fernet-keys\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.258032 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-config-data\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.258131 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-combined-ca-bundle\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.258169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkdxd\" (UniqueName: \"kubernetes.io/projected/6f205373-c895-453e-90bd-9c323f6c90ac-kube-api-access-vkdxd\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.360626 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-config-data\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.360734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-combined-ca-bundle\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.360758 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkdxd\" (UniqueName: \"kubernetes.io/projected/6f205373-c895-453e-90bd-9c323f6c90ac-kube-api-access-vkdxd\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.360867 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-fernet-keys\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.368218 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-fernet-keys\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.368203 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-combined-ca-bundle\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.378856 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-config-data\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.381721 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkdxd\" (UniqueName: \"kubernetes.io/projected/6f205373-c895-453e-90bd-9c323f6c90ac-kube-api-access-vkdxd\") pod \"keystone-cron-29320741-x298c\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:00 crc kubenswrapper[4799]: I0930 15:01:00.518009 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:01 crc kubenswrapper[4799]: I0930 15:01:01.020976 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29320741-x298c"] Sep 30 15:01:01 crc kubenswrapper[4799]: I0930 15:01:01.284842 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320741-x298c" event={"ID":"6f205373-c895-453e-90bd-9c323f6c90ac","Type":"ContainerStarted","Data":"1c9d4edd8df76058be7f3ef8de2db5f63ebdd17456b57915f2e97fc38eeab3b3"} Sep 30 15:01:01 crc kubenswrapper[4799]: I0930 15:01:01.285226 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320741-x298c" event={"ID":"6f205373-c895-453e-90bd-9c323f6c90ac","Type":"ContainerStarted","Data":"b08bf58a159f814b40d3f8321288fa2ce26ac1e0383b9442babdf20d72b24c51"} Sep 30 15:01:01 crc kubenswrapper[4799]: I0930 15:01:01.312280 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29320741-x298c" podStartSLOduration=1.312249974 podStartE2EDuration="1.312249974s" podCreationTimestamp="2025-09-30 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:01:01.306000721 +0000 UTC m=+2483.389601158" watchObservedRunningTime="2025-09-30 15:01:01.312249974 +0000 UTC m=+2483.395850411" Sep 30 15:01:05 crc kubenswrapper[4799]: I0930 15:01:05.336866 4799 generic.go:334] "Generic (PLEG): container finished" podID="6f205373-c895-453e-90bd-9c323f6c90ac" containerID="1c9d4edd8df76058be7f3ef8de2db5f63ebdd17456b57915f2e97fc38eeab3b3" exitCode=0 Sep 30 15:01:05 crc kubenswrapper[4799]: I0930 15:01:05.337016 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320741-x298c" event={"ID":"6f205373-c895-453e-90bd-9c323f6c90ac","Type":"ContainerDied","Data":"1c9d4edd8df76058be7f3ef8de2db5f63ebdd17456b57915f2e97fc38eeab3b3"} Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.676703 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.803490 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-config-data\") pod \"6f205373-c895-453e-90bd-9c323f6c90ac\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.803952 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkdxd\" (UniqueName: \"kubernetes.io/projected/6f205373-c895-453e-90bd-9c323f6c90ac-kube-api-access-vkdxd\") pod \"6f205373-c895-453e-90bd-9c323f6c90ac\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.804220 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-combined-ca-bundle\") pod \"6f205373-c895-453e-90bd-9c323f6c90ac\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.804369 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-fernet-keys\") pod \"6f205373-c895-453e-90bd-9c323f6c90ac\" (UID: \"6f205373-c895-453e-90bd-9c323f6c90ac\") " Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.810158 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6f205373-c895-453e-90bd-9c323f6c90ac" (UID: "6f205373-c895-453e-90bd-9c323f6c90ac"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.810264 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f205373-c895-453e-90bd-9c323f6c90ac-kube-api-access-vkdxd" (OuterVolumeSpecName: "kube-api-access-vkdxd") pod "6f205373-c895-453e-90bd-9c323f6c90ac" (UID: "6f205373-c895-453e-90bd-9c323f6c90ac"). InnerVolumeSpecName "kube-api-access-vkdxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.838395 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f205373-c895-453e-90bd-9c323f6c90ac" (UID: "6f205373-c895-453e-90bd-9c323f6c90ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.864941 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-config-data" (OuterVolumeSpecName: "config-data") pod "6f205373-c895-453e-90bd-9c323f6c90ac" (UID: "6f205373-c895-453e-90bd-9c323f6c90ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.907951 4799 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-fernet-keys\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.907992 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.908006 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkdxd\" (UniqueName: \"kubernetes.io/projected/6f205373-c895-453e-90bd-9c323f6c90ac-kube-api-access-vkdxd\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:06 crc kubenswrapper[4799]: I0930 15:01:06.908019 4799 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f205373-c895-453e-90bd-9c323f6c90ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 15:01:07 crc kubenswrapper[4799]: I0930 15:01:07.358135 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29320741-x298c" event={"ID":"6f205373-c895-453e-90bd-9c323f6c90ac","Type":"ContainerDied","Data":"b08bf58a159f814b40d3f8321288fa2ce26ac1e0383b9442babdf20d72b24c51"} Sep 30 15:01:07 crc kubenswrapper[4799]: I0930 15:01:07.358186 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b08bf58a159f814b40d3f8321288fa2ce26ac1e0383b9442babdf20d72b24c51" Sep 30 15:01:07 crc kubenswrapper[4799]: I0930 15:01:07.358222 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29320741-x298c" Sep 30 15:01:07 crc kubenswrapper[4799]: I0930 15:01:07.503214 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:01:07 crc kubenswrapper[4799]: E0930 15:01:07.503503 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:01:20 crc kubenswrapper[4799]: I0930 15:01:20.503543 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:01:20 crc kubenswrapper[4799]: E0930 15:01:20.504308 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:01:33 crc kubenswrapper[4799]: I0930 15:01:33.504490 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:01:33 crc kubenswrapper[4799]: E0930 15:01:33.506918 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:01:44 crc kubenswrapper[4799]: I0930 15:01:44.503570 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:01:44 crc kubenswrapper[4799]: E0930 15:01:44.504441 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:01:56 crc kubenswrapper[4799]: I0930 15:01:56.503705 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:01:56 crc kubenswrapper[4799]: E0930 15:01:56.504343 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:02:10 crc kubenswrapper[4799]: I0930 15:02:10.509181 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:02:11 crc kubenswrapper[4799]: I0930 15:02:11.002798 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"8ab6660d9eab2d41fdd2de146c8c871d31c86110d25066571e7ae3aa2faee410"} Sep 30 15:04:24 crc kubenswrapper[4799]: I0930 15:04:24.351629 4799 generic.go:334] "Generic (PLEG): container finished" podID="43ae8f90-c778-47e6-8c69-8cdd2b4fa949" containerID="9d069c373284b5539ad2a9103fbee9e88bb6ea9ef45fb434baba875b8fdf2c48" exitCode=0 Sep 30 15:04:24 crc kubenswrapper[4799]: I0930 15:04:24.351890 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" event={"ID":"43ae8f90-c778-47e6-8c69-8cdd2b4fa949","Type":"ContainerDied","Data":"9d069c373284b5539ad2a9103fbee9e88bb6ea9ef45fb434baba875b8fdf2c48"} Sep 30 15:04:25 crc kubenswrapper[4799]: I0930 15:04:25.854145 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 15:04:25 crc kubenswrapper[4799]: I0930 15:04:25.967250 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-combined-ca-bundle\") pod \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " Sep 30 15:04:25 crc kubenswrapper[4799]: I0930 15:04:25.967327 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-ssh-key\") pod \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " Sep 30 15:04:25 crc kubenswrapper[4799]: I0930 15:04:25.967439 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-secret-0\") pod \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " Sep 30 15:04:25 crc kubenswrapper[4799]: I0930 15:04:25.967631 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-inventory\") pod \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " Sep 30 15:04:25 crc kubenswrapper[4799]: I0930 15:04:25.968494 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw9vt\" (UniqueName: \"kubernetes.io/projected/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-kube-api-access-mw9vt\") pod \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\" (UID: \"43ae8f90-c778-47e6-8c69-8cdd2b4fa949\") " Sep 30 15:04:25 crc kubenswrapper[4799]: I0930 15:04:25.976244 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-kube-api-access-mw9vt" (OuterVolumeSpecName: "kube-api-access-mw9vt") pod "43ae8f90-c778-47e6-8c69-8cdd2b4fa949" (UID: "43ae8f90-c778-47e6-8c69-8cdd2b4fa949"). InnerVolumeSpecName "kube-api-access-mw9vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:04:25 crc kubenswrapper[4799]: I0930 15:04:25.976675 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "43ae8f90-c778-47e6-8c69-8cdd2b4fa949" (UID: "43ae8f90-c778-47e6-8c69-8cdd2b4fa949"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.005901 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43ae8f90-c778-47e6-8c69-8cdd2b4fa949" (UID: "43ae8f90-c778-47e6-8c69-8cdd2b4fa949"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.006087 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "43ae8f90-c778-47e6-8c69-8cdd2b4fa949" (UID: "43ae8f90-c778-47e6-8c69-8cdd2b4fa949"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.009323 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-inventory" (OuterVolumeSpecName: "inventory") pod "43ae8f90-c778-47e6-8c69-8cdd2b4fa949" (UID: "43ae8f90-c778-47e6-8c69-8cdd2b4fa949"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.071633 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.071687 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.071699 4799 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.071710 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.071725 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw9vt\" (UniqueName: \"kubernetes.io/projected/43ae8f90-c778-47e6-8c69-8cdd2b4fa949-kube-api-access-mw9vt\") on node \"crc\" DevicePath \"\"" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.375422 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" event={"ID":"43ae8f90-c778-47e6-8c69-8cdd2b4fa949","Type":"ContainerDied","Data":"b980eafbbfeca39d1ab89bf46cb6727a5a74ec91f67ac8b4dd6198f1de057fa4"} Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.375486 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b980eafbbfeca39d1ab89bf46cb6727a5a74ec91f67ac8b4dd6198f1de057fa4" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.375557 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.641565 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx"] Sep 30 15:04:26 crc kubenswrapper[4799]: E0930 15:04:26.642810 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43ae8f90-c778-47e6-8c69-8cdd2b4fa949" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.642848 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="43ae8f90-c778-47e6-8c69-8cdd2b4fa949" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 15:04:26 crc kubenswrapper[4799]: E0930 15:04:26.642928 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f205373-c895-453e-90bd-9c323f6c90ac" containerName="keystone-cron" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.642939 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f205373-c895-453e-90bd-9c323f6c90ac" containerName="keystone-cron" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.643178 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f205373-c895-453e-90bd-9c323f6c90ac" containerName="keystone-cron" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.643202 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="43ae8f90-c778-47e6-8c69-8cdd2b4fa949" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.644207 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.646768 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.648037 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.648750 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.649921 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.651444 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.651715 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.654559 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.668965 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx"] Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.689985 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjgmm\" (UniqueName: \"kubernetes.io/projected/2da7cba0-9bbc-47c7-99b0-f204928886fa-kube-api-access-qjgmm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.690085 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.690121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.690147 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.690186 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.690239 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.690346 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.690371 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.690584 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.792956 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.793100 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjgmm\" (UniqueName: \"kubernetes.io/projected/2da7cba0-9bbc-47c7-99b0-f204928886fa-kube-api-access-qjgmm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.793136 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.793166 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.793193 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.793226 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.793263 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.793500 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.793539 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.794460 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.799836 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.800065 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.801136 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.802347 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.802788 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.807220 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.812428 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.820197 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjgmm\" (UniqueName: \"kubernetes.io/projected/2da7cba0-9bbc-47c7-99b0-f204928886fa-kube-api-access-qjgmm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tlstx\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:26 crc kubenswrapper[4799]: I0930 15:04:26.963704 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:04:27 crc kubenswrapper[4799]: I0930 15:04:27.530750 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx"] Sep 30 15:04:27 crc kubenswrapper[4799]: I0930 15:04:27.545734 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 15:04:28 crc kubenswrapper[4799]: I0930 15:04:28.398306 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" event={"ID":"2da7cba0-9bbc-47c7-99b0-f204928886fa","Type":"ContainerStarted","Data":"3f4ee7a3ac8bbd5c6b28cecd6d03a66d61e719121e04c450f76ed0dd75c5d916"} Sep 30 15:04:28 crc kubenswrapper[4799]: I0930 15:04:28.398671 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" event={"ID":"2da7cba0-9bbc-47c7-99b0-f204928886fa","Type":"ContainerStarted","Data":"b2d1f15412cf4c7767feccea56c10fd438e7e2fa432e6a0b194f0b79db1bed41"} Sep 30 15:04:28 crc kubenswrapper[4799]: I0930 15:04:28.430166 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" podStartSLOduration=2.243455001 podStartE2EDuration="2.430137881s" podCreationTimestamp="2025-09-30 15:04:26 +0000 UTC" firstStartedPulling="2025-09-30 15:04:27.545449664 +0000 UTC m=+2689.629050091" lastFinishedPulling="2025-09-30 15:04:27.732132534 +0000 UTC m=+2689.815732971" observedRunningTime="2025-09-30 15:04:28.428515753 +0000 UTC m=+2690.512116190" watchObservedRunningTime="2025-09-30 15:04:28.430137881 +0000 UTC m=+2690.513738308" Sep 30 15:04:29 crc kubenswrapper[4799]: I0930 15:04:29.648646 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:04:29 crc kubenswrapper[4799]: I0930 15:04:29.649131 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:04:59 crc kubenswrapper[4799]: I0930 15:04:59.648678 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:04:59 crc kubenswrapper[4799]: I0930 15:04:59.649275 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.769116 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rnzqz"] Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.776449 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.789757 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnzqz"] Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.830968 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8lrc\" (UniqueName: \"kubernetes.io/projected/95e3f792-4726-4292-9aa3-a94919534022-kube-api-access-z8lrc\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.831596 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e3f792-4726-4292-9aa3-a94919534022-catalog-content\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.831637 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e3f792-4726-4292-9aa3-a94919534022-utilities\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.933950 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8lrc\" (UniqueName: \"kubernetes.io/projected/95e3f792-4726-4292-9aa3-a94919534022-kube-api-access-z8lrc\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.934047 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e3f792-4726-4292-9aa3-a94919534022-catalog-content\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.934082 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e3f792-4726-4292-9aa3-a94919534022-utilities\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.934761 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e3f792-4726-4292-9aa3-a94919534022-utilities\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.934968 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e3f792-4726-4292-9aa3-a94919534022-catalog-content\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:08 crc kubenswrapper[4799]: I0930 15:05:08.968320 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8lrc\" (UniqueName: \"kubernetes.io/projected/95e3f792-4726-4292-9aa3-a94919534022-kube-api-access-z8lrc\") pod \"community-operators-rnzqz\" (UID: \"95e3f792-4726-4292-9aa3-a94919534022\") " pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:09 crc kubenswrapper[4799]: I0930 15:05:09.102800 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:09 crc kubenswrapper[4799]: I0930 15:05:09.759796 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnzqz"] Sep 30 15:05:09 crc kubenswrapper[4799]: I0930 15:05:09.837025 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnzqz" event={"ID":"95e3f792-4726-4292-9aa3-a94919534022","Type":"ContainerStarted","Data":"d954369187bba2b322350fdda6a6cf378f13ea177c41e3166f24a00484df8493"} Sep 30 15:05:10 crc kubenswrapper[4799]: I0930 15:05:10.857037 4799 generic.go:334] "Generic (PLEG): container finished" podID="95e3f792-4726-4292-9aa3-a94919534022" containerID="2dfebcd5b0783d2788e4a32a0a51b7730bd54cd3d8c74faf324ac6b52d244825" exitCode=0 Sep 30 15:05:10 crc kubenswrapper[4799]: I0930 15:05:10.857395 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnzqz" event={"ID":"95e3f792-4726-4292-9aa3-a94919534022","Type":"ContainerDied","Data":"2dfebcd5b0783d2788e4a32a0a51b7730bd54cd3d8c74faf324ac6b52d244825"} Sep 30 15:05:15 crc kubenswrapper[4799]: I0930 15:05:15.910459 4799 generic.go:334] "Generic (PLEG): container finished" podID="95e3f792-4726-4292-9aa3-a94919534022" containerID="eff8cc4136970ace6041c73c7ed4741698c9a29bd7de70d34e52015f987699a0" exitCode=0 Sep 30 15:05:15 crc kubenswrapper[4799]: I0930 15:05:15.910579 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnzqz" event={"ID":"95e3f792-4726-4292-9aa3-a94919534022","Type":"ContainerDied","Data":"eff8cc4136970ace6041c73c7ed4741698c9a29bd7de70d34e52015f987699a0"} Sep 30 15:05:16 crc kubenswrapper[4799]: I0930 15:05:16.928987 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnzqz" event={"ID":"95e3f792-4726-4292-9aa3-a94919534022","Type":"ContainerStarted","Data":"6cdeba221cb143b1fa5384e233b9e4c23236a71fe6c268f988aca04edbb8392e"} Sep 30 15:05:16 crc kubenswrapper[4799]: I0930 15:05:16.956830 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rnzqz" podStartSLOduration=3.346724663 podStartE2EDuration="8.956798672s" podCreationTimestamp="2025-09-30 15:05:08 +0000 UTC" firstStartedPulling="2025-09-30 15:05:10.860258144 +0000 UTC m=+2732.943858571" lastFinishedPulling="2025-09-30 15:05:16.470332153 +0000 UTC m=+2738.553932580" observedRunningTime="2025-09-30 15:05:16.950802756 +0000 UTC m=+2739.034403203" watchObservedRunningTime="2025-09-30 15:05:16.956798672 +0000 UTC m=+2739.040399099" Sep 30 15:05:19 crc kubenswrapper[4799]: I0930 15:05:19.103317 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:19 crc kubenswrapper[4799]: I0930 15:05:19.103700 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:19 crc kubenswrapper[4799]: I0930 15:05:19.156885 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.741083 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m7r68"] Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.744032 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.760440 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m7r68"] Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.788679 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-utilities\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.789284 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn4cn\" (UniqueName: \"kubernetes.io/projected/56db947e-810d-4bce-aefd-1d7e245299ff-kube-api-access-nn4cn\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.789450 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-catalog-content\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.891115 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-utilities\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.891316 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn4cn\" (UniqueName: \"kubernetes.io/projected/56db947e-810d-4bce-aefd-1d7e245299ff-kube-api-access-nn4cn\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.891362 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-catalog-content\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.891814 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-utilities\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.891970 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-catalog-content\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:23 crc kubenswrapper[4799]: I0930 15:05:23.922088 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn4cn\" (UniqueName: \"kubernetes.io/projected/56db947e-810d-4bce-aefd-1d7e245299ff-kube-api-access-nn4cn\") pod \"redhat-operators-m7r68\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:24 crc kubenswrapper[4799]: I0930 15:05:24.071933 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:24 crc kubenswrapper[4799]: I0930 15:05:24.603895 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m7r68"] Sep 30 15:05:25 crc kubenswrapper[4799]: I0930 15:05:25.013177 4799 generic.go:334] "Generic (PLEG): container finished" podID="56db947e-810d-4bce-aefd-1d7e245299ff" containerID="d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d" exitCode=0 Sep 30 15:05:25 crc kubenswrapper[4799]: I0930 15:05:25.013451 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7r68" event={"ID":"56db947e-810d-4bce-aefd-1d7e245299ff","Type":"ContainerDied","Data":"d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d"} Sep 30 15:05:25 crc kubenswrapper[4799]: I0930 15:05:25.013483 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7r68" event={"ID":"56db947e-810d-4bce-aefd-1d7e245299ff","Type":"ContainerStarted","Data":"b44a29a45d2b1850618b5c30a797f4988a6ef1085208637c8db3d4c7bde35460"} Sep 30 15:05:26 crc kubenswrapper[4799]: I0930 15:05:26.029859 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7r68" event={"ID":"56db947e-810d-4bce-aefd-1d7e245299ff","Type":"ContainerStarted","Data":"92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d"} Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.157791 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rnzqz" Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.276238 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnzqz"] Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.345839 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nbzj"] Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.346128 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6nbzj" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerName="registry-server" containerID="cri-o://98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f" gracePeriod=2 Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.649559 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.649923 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.649980 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.650902 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ab6660d9eab2d41fdd2de146c8c871d31c86110d25066571e7ae3aa2faee410"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:05:29 crc kubenswrapper[4799]: I0930 15:05:29.650953 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://8ab6660d9eab2d41fdd2de146c8c871d31c86110d25066571e7ae3aa2faee410" gracePeriod=600 Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.010566 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nbzj" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.036795 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-utilities\") pod \"7635e2e9-575c-43ff-894f-e8bd89dacb28\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.036889 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-catalog-content\") pod \"7635e2e9-575c-43ff-894f-e8bd89dacb28\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.037019 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pghhq\" (UniqueName: \"kubernetes.io/projected/7635e2e9-575c-43ff-894f-e8bd89dacb28-kube-api-access-pghhq\") pod \"7635e2e9-575c-43ff-894f-e8bd89dacb28\" (UID: \"7635e2e9-575c-43ff-894f-e8bd89dacb28\") " Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.038437 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-utilities" (OuterVolumeSpecName: "utilities") pod "7635e2e9-575c-43ff-894f-e8bd89dacb28" (UID: "7635e2e9-575c-43ff-894f-e8bd89dacb28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.045954 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7635e2e9-575c-43ff-894f-e8bd89dacb28-kube-api-access-pghhq" (OuterVolumeSpecName: "kube-api-access-pghhq") pod "7635e2e9-575c-43ff-894f-e8bd89dacb28" (UID: "7635e2e9-575c-43ff-894f-e8bd89dacb28"). InnerVolumeSpecName "kube-api-access-pghhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.098884 4799 generic.go:334] "Generic (PLEG): container finished" podID="56db947e-810d-4bce-aefd-1d7e245299ff" containerID="92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d" exitCode=0 Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.098983 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7r68" event={"ID":"56db947e-810d-4bce-aefd-1d7e245299ff","Type":"ContainerDied","Data":"92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d"} Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.124063 4799 generic.go:334] "Generic (PLEG): container finished" podID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerID="98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f" exitCode=0 Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.124279 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nbzj" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.124857 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nbzj" event={"ID":"7635e2e9-575c-43ff-894f-e8bd89dacb28","Type":"ContainerDied","Data":"98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f"} Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.124905 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nbzj" event={"ID":"7635e2e9-575c-43ff-894f-e8bd89dacb28","Type":"ContainerDied","Data":"7ec2838cb8173e1c7b4038a2d8fbdd2234d70e55d6c268fd5ef476b0b55b068a"} Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.124929 4799 scope.go:117] "RemoveContainer" containerID="98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.139635 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pghhq\" (UniqueName: \"kubernetes.io/projected/7635e2e9-575c-43ff-894f-e8bd89dacb28-kube-api-access-pghhq\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.139722 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.142058 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="8ab6660d9eab2d41fdd2de146c8c871d31c86110d25066571e7ae3aa2faee410" exitCode=0 Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.142837 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"8ab6660d9eab2d41fdd2de146c8c871d31c86110d25066571e7ae3aa2faee410"} Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.219632 4799 scope.go:117] "RemoveContainer" containerID="dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.253517 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7635e2e9-575c-43ff-894f-e8bd89dacb28" (UID: "7635e2e9-575c-43ff-894f-e8bd89dacb28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.290830 4799 scope.go:117] "RemoveContainer" containerID="73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.350234 4799 scope.go:117] "RemoveContainer" containerID="98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f" Sep 30 15:05:30 crc kubenswrapper[4799]: E0930 15:05:30.353882 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f\": container with ID starting with 98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f not found: ID does not exist" containerID="98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.353952 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f"} err="failed to get container status \"98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f\": rpc error: code = NotFound desc = could not find container \"98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f\": container with ID starting with 98e500fccc3d5482c630cd435e0b4246e3225d49621310bd236741ea16c8458f not found: ID does not exist" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.353993 4799 scope.go:117] "RemoveContainer" containerID="dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.354443 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7635e2e9-575c-43ff-894f-e8bd89dacb28-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:30 crc kubenswrapper[4799]: E0930 15:05:30.356610 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801\": container with ID starting with dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801 not found: ID does not exist" containerID="dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.356701 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801"} err="failed to get container status \"dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801\": rpc error: code = NotFound desc = could not find container \"dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801\": container with ID starting with dc2c6740417bc47f88bd51d33c28bd64f190c30eb53043f080b28f6b0ac44801 not found: ID does not exist" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.356735 4799 scope.go:117] "RemoveContainer" containerID="73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64" Sep 30 15:05:30 crc kubenswrapper[4799]: E0930 15:05:30.360036 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64\": container with ID starting with 73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64 not found: ID does not exist" containerID="73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.360097 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64"} err="failed to get container status \"73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64\": rpc error: code = NotFound desc = could not find container \"73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64\": container with ID starting with 73bee2749d143b8c4793a8f76b4f6bf96e65616bf87cc26a8d42a7fa2ad67b64 not found: ID does not exist" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.360135 4799 scope.go:117] "RemoveContainer" containerID="d4ee3fa1a82bb172be84f13f467d15bc48f4e88f8d0f1546c891f5c67b7c2976" Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.503384 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nbzj"] Sep 30 15:05:30 crc kubenswrapper[4799]: I0930 15:05:30.534718 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6nbzj"] Sep 30 15:05:31 crc kubenswrapper[4799]: I0930 15:05:31.156626 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7r68" event={"ID":"56db947e-810d-4bce-aefd-1d7e245299ff","Type":"ContainerStarted","Data":"4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f"} Sep 30 15:05:31 crc kubenswrapper[4799]: I0930 15:05:31.161986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53"} Sep 30 15:05:31 crc kubenswrapper[4799]: I0930 15:05:31.190929 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m7r68" podStartSLOduration=2.466406347 podStartE2EDuration="8.190883994s" podCreationTimestamp="2025-09-30 15:05:23 +0000 UTC" firstStartedPulling="2025-09-30 15:05:25.015008407 +0000 UTC m=+2747.098608834" lastFinishedPulling="2025-09-30 15:05:30.739486054 +0000 UTC m=+2752.823086481" observedRunningTime="2025-09-30 15:05:31.189200555 +0000 UTC m=+2753.272800992" watchObservedRunningTime="2025-09-30 15:05:31.190883994 +0000 UTC m=+2753.274484421" Sep 30 15:05:32 crc kubenswrapper[4799]: I0930 15:05:32.517300 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" path="/var/lib/kubelet/pods/7635e2e9-575c-43ff-894f-e8bd89dacb28/volumes" Sep 30 15:05:34 crc kubenswrapper[4799]: I0930 15:05:34.072302 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:34 crc kubenswrapper[4799]: I0930 15:05:34.072633 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:35 crc kubenswrapper[4799]: I0930 15:05:35.128687 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m7r68" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="registry-server" probeResult="failure" output=< Sep 30 15:05:35 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:05:35 crc kubenswrapper[4799]: > Sep 30 15:05:45 crc kubenswrapper[4799]: I0930 15:05:45.120437 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m7r68" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="registry-server" probeResult="failure" output=< Sep 30 15:05:45 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:05:45 crc kubenswrapper[4799]: > Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.369442 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k9rmd"] Sep 30 15:05:48 crc kubenswrapper[4799]: E0930 15:05:48.370394 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerName="extract-utilities" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.370411 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerName="extract-utilities" Sep 30 15:05:48 crc kubenswrapper[4799]: E0930 15:05:48.370428 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerName="extract-content" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.370436 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerName="extract-content" Sep 30 15:05:48 crc kubenswrapper[4799]: E0930 15:05:48.370484 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerName="registry-server" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.370493 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerName="registry-server" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.370755 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7635e2e9-575c-43ff-894f-e8bd89dacb28" containerName="registry-server" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.372854 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.397183 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9rmd"] Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.536927 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-utilities\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.537017 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tv6x\" (UniqueName: \"kubernetes.io/projected/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-kube-api-access-2tv6x\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.537057 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-catalog-content\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.639175 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-utilities\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.639252 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tv6x\" (UniqueName: \"kubernetes.io/projected/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-kube-api-access-2tv6x\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.639296 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-catalog-content\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.639733 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-utilities\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.640414 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-catalog-content\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.684971 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tv6x\" (UniqueName: \"kubernetes.io/projected/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-kube-api-access-2tv6x\") pod \"redhat-marketplace-k9rmd\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:48 crc kubenswrapper[4799]: I0930 15:05:48.703180 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:49 crc kubenswrapper[4799]: I0930 15:05:49.249697 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9rmd"] Sep 30 15:05:49 crc kubenswrapper[4799]: I0930 15:05:49.362517 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9rmd" event={"ID":"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a","Type":"ContainerStarted","Data":"a0121565fe43bb6fd87273a87d06f6cd48da9ab90cdea0382f3dec85a8631e98"} Sep 30 15:05:50 crc kubenswrapper[4799]: I0930 15:05:50.373856 4799 generic.go:334] "Generic (PLEG): container finished" podID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerID="6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b" exitCode=0 Sep 30 15:05:50 crc kubenswrapper[4799]: I0930 15:05:50.374040 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9rmd" event={"ID":"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a","Type":"ContainerDied","Data":"6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b"} Sep 30 15:05:51 crc kubenswrapper[4799]: I0930 15:05:51.387406 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9rmd" event={"ID":"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a","Type":"ContainerStarted","Data":"0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213"} Sep 30 15:05:52 crc kubenswrapper[4799]: I0930 15:05:52.398864 4799 generic.go:334] "Generic (PLEG): container finished" podID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerID="0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213" exitCode=0 Sep 30 15:05:52 crc kubenswrapper[4799]: I0930 15:05:52.398921 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9rmd" event={"ID":"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a","Type":"ContainerDied","Data":"0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213"} Sep 30 15:05:53 crc kubenswrapper[4799]: I0930 15:05:53.432500 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9rmd" event={"ID":"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a","Type":"ContainerStarted","Data":"d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77"} Sep 30 15:05:53 crc kubenswrapper[4799]: I0930 15:05:53.464049 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k9rmd" podStartSLOduration=2.904845598 podStartE2EDuration="5.464020772s" podCreationTimestamp="2025-09-30 15:05:48 +0000 UTC" firstStartedPulling="2025-09-30 15:05:50.376309025 +0000 UTC m=+2772.459909462" lastFinishedPulling="2025-09-30 15:05:52.935484209 +0000 UTC m=+2775.019084636" observedRunningTime="2025-09-30 15:05:53.459307573 +0000 UTC m=+2775.542908010" watchObservedRunningTime="2025-09-30 15:05:53.464020772 +0000 UTC m=+2775.547621209" Sep 30 15:05:54 crc kubenswrapper[4799]: I0930 15:05:54.122349 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:54 crc kubenswrapper[4799]: I0930 15:05:54.173140 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:55 crc kubenswrapper[4799]: I0930 15:05:55.739101 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m7r68"] Sep 30 15:05:55 crc kubenswrapper[4799]: I0930 15:05:55.739841 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m7r68" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="registry-server" containerID="cri-o://4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f" gracePeriod=2 Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.186363 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.315233 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn4cn\" (UniqueName: \"kubernetes.io/projected/56db947e-810d-4bce-aefd-1d7e245299ff-kube-api-access-nn4cn\") pod \"56db947e-810d-4bce-aefd-1d7e245299ff\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.315674 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-catalog-content\") pod \"56db947e-810d-4bce-aefd-1d7e245299ff\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.315736 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-utilities\") pod \"56db947e-810d-4bce-aefd-1d7e245299ff\" (UID: \"56db947e-810d-4bce-aefd-1d7e245299ff\") " Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.316458 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-utilities" (OuterVolumeSpecName: "utilities") pod "56db947e-810d-4bce-aefd-1d7e245299ff" (UID: "56db947e-810d-4bce-aefd-1d7e245299ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.325137 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56db947e-810d-4bce-aefd-1d7e245299ff-kube-api-access-nn4cn" (OuterVolumeSpecName: "kube-api-access-nn4cn") pod "56db947e-810d-4bce-aefd-1d7e245299ff" (UID: "56db947e-810d-4bce-aefd-1d7e245299ff"). InnerVolumeSpecName "kube-api-access-nn4cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.416795 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56db947e-810d-4bce-aefd-1d7e245299ff" (UID: "56db947e-810d-4bce-aefd-1d7e245299ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.418404 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn4cn\" (UniqueName: \"kubernetes.io/projected/56db947e-810d-4bce-aefd-1d7e245299ff-kube-api-access-nn4cn\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.418451 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.418461 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56db947e-810d-4bce-aefd-1d7e245299ff-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.468045 4799 generic.go:334] "Generic (PLEG): container finished" podID="56db947e-810d-4bce-aefd-1d7e245299ff" containerID="4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f" exitCode=0 Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.468108 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7r68" event={"ID":"56db947e-810d-4bce-aefd-1d7e245299ff","Type":"ContainerDied","Data":"4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f"} Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.468154 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7r68" event={"ID":"56db947e-810d-4bce-aefd-1d7e245299ff","Type":"ContainerDied","Data":"b44a29a45d2b1850618b5c30a797f4988a6ef1085208637c8db3d4c7bde35460"} Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.468178 4799 scope.go:117] "RemoveContainer" containerID="4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.468185 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7r68" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.495354 4799 scope.go:117] "RemoveContainer" containerID="92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.562927 4799 scope.go:117] "RemoveContainer" containerID="d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.573273 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m7r68"] Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.598546 4799 scope.go:117] "RemoveContainer" containerID="4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f" Sep 30 15:05:56 crc kubenswrapper[4799]: E0930 15:05:56.599360 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f\": container with ID starting with 4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f not found: ID does not exist" containerID="4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.599497 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f"} err="failed to get container status \"4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f\": rpc error: code = NotFound desc = could not find container \"4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f\": container with ID starting with 4bf21b9ee527ed76304164d9c1bf673789901beecb92d156c34487cb8504129f not found: ID does not exist" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.599608 4799 scope.go:117] "RemoveContainer" containerID="92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d" Sep 30 15:05:56 crc kubenswrapper[4799]: E0930 15:05:56.600638 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d\": container with ID starting with 92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d not found: ID does not exist" containerID="92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.600709 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d"} err="failed to get container status \"92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d\": rpc error: code = NotFound desc = could not find container \"92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d\": container with ID starting with 92c92924d9148628b72e403e1768cccddb9ff8dc78d8bf8f56ad7fdee6360b5d not found: ID does not exist" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.600748 4799 scope.go:117] "RemoveContainer" containerID="d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d" Sep 30 15:05:56 crc kubenswrapper[4799]: E0930 15:05:56.601797 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d\": container with ID starting with d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d not found: ID does not exist" containerID="d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.602211 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d"} err="failed to get container status \"d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d\": rpc error: code = NotFound desc = could not find container \"d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d\": container with ID starting with d5da0c315bfb6d5747da432923dd9e217019d6cbce3c0ee98020329b21a3d85d not found: ID does not exist" Sep 30 15:05:56 crc kubenswrapper[4799]: I0930 15:05:56.606523 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m7r68"] Sep 30 15:05:58 crc kubenswrapper[4799]: I0930 15:05:58.519413 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" path="/var/lib/kubelet/pods/56db947e-810d-4bce-aefd-1d7e245299ff/volumes" Sep 30 15:05:58 crc kubenswrapper[4799]: I0930 15:05:58.705588 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:58 crc kubenswrapper[4799]: I0930 15:05:58.707857 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:58 crc kubenswrapper[4799]: I0930 15:05:58.771442 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:59 crc kubenswrapper[4799]: I0930 15:05:59.571170 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:05:59 crc kubenswrapper[4799]: I0930 15:05:59.937288 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9rmd"] Sep 30 15:06:01 crc kubenswrapper[4799]: I0930 15:06:01.522910 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k9rmd" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerName="registry-server" containerID="cri-o://d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77" gracePeriod=2 Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.002200 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.079328 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-utilities\") pod \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.079571 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tv6x\" (UniqueName: \"kubernetes.io/projected/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-kube-api-access-2tv6x\") pod \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.079977 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-catalog-content\") pod \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\" (UID: \"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a\") " Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.082485 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-utilities" (OuterVolumeSpecName: "utilities") pod "a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" (UID: "a8745c7a-ec3a-4bc7-9ef1-7de335cf600a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.090281 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-kube-api-access-2tv6x" (OuterVolumeSpecName: "kube-api-access-2tv6x") pod "a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" (UID: "a8745c7a-ec3a-4bc7-9ef1-7de335cf600a"). InnerVolumeSpecName "kube-api-access-2tv6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.102597 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" (UID: "a8745c7a-ec3a-4bc7-9ef1-7de335cf600a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.182303 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.182368 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.182383 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tv6x\" (UniqueName: \"kubernetes.io/projected/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a-kube-api-access-2tv6x\") on node \"crc\" DevicePath \"\"" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.544225 4799 generic.go:334] "Generic (PLEG): container finished" podID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerID="d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77" exitCode=0 Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.544280 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9rmd" event={"ID":"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a","Type":"ContainerDied","Data":"d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77"} Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.544317 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k9rmd" event={"ID":"a8745c7a-ec3a-4bc7-9ef1-7de335cf600a","Type":"ContainerDied","Data":"a0121565fe43bb6fd87273a87d06f6cd48da9ab90cdea0382f3dec85a8631e98"} Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.544338 4799 scope.go:117] "RemoveContainer" containerID="d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.544493 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k9rmd" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.583439 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9rmd"] Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.587734 4799 scope.go:117] "RemoveContainer" containerID="0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.592401 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k9rmd"] Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.621321 4799 scope.go:117] "RemoveContainer" containerID="6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.678851 4799 scope.go:117] "RemoveContainer" containerID="d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77" Sep 30 15:06:02 crc kubenswrapper[4799]: E0930 15:06:02.679534 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77\": container with ID starting with d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77 not found: ID does not exist" containerID="d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.679599 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77"} err="failed to get container status \"d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77\": rpc error: code = NotFound desc = could not find container \"d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77\": container with ID starting with d0401d94eebaa6c165d105f61948f32678caab881082d149455fcde9cc76ad77 not found: ID does not exist" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.679635 4799 scope.go:117] "RemoveContainer" containerID="0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213" Sep 30 15:06:02 crc kubenswrapper[4799]: E0930 15:06:02.680252 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213\": container with ID starting with 0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213 not found: ID does not exist" containerID="0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.680269 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213"} err="failed to get container status \"0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213\": rpc error: code = NotFound desc = could not find container \"0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213\": container with ID starting with 0274d01d617a712b353c5b4d5eff8d8aa2d24a0753f561242e97c45426280213 not found: ID does not exist" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.680288 4799 scope.go:117] "RemoveContainer" containerID="6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b" Sep 30 15:06:02 crc kubenswrapper[4799]: E0930 15:06:02.680797 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b\": container with ID starting with 6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b not found: ID does not exist" containerID="6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b" Sep 30 15:06:02 crc kubenswrapper[4799]: I0930 15:06:02.680823 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b"} err="failed to get container status \"6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b\": rpc error: code = NotFound desc = could not find container \"6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b\": container with ID starting with 6d99325d7286f6b7ae62d2aec28b3ee4e59e596263b4ad9e8c5e46960607933b not found: ID does not exist" Sep 30 15:06:04 crc kubenswrapper[4799]: I0930 15:06:04.520458 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" path="/var/lib/kubelet/pods/a8745c7a-ec3a-4bc7-9ef1-7de335cf600a/volumes" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.230072 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6f568"] Sep 30 15:07:29 crc kubenswrapper[4799]: E0930 15:07:29.233685 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerName="registry-server" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.233733 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerName="registry-server" Sep 30 15:07:29 crc kubenswrapper[4799]: E0930 15:07:29.233745 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerName="extract-content" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.233751 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerName="extract-content" Sep 30 15:07:29 crc kubenswrapper[4799]: E0930 15:07:29.233773 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="registry-server" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.233778 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="registry-server" Sep 30 15:07:29 crc kubenswrapper[4799]: E0930 15:07:29.233804 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="extract-utilities" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.233813 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="extract-utilities" Sep 30 15:07:29 crc kubenswrapper[4799]: E0930 15:07:29.233826 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="extract-content" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.233834 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="extract-content" Sep 30 15:07:29 crc kubenswrapper[4799]: E0930 15:07:29.233848 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerName="extract-utilities" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.233883 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerName="extract-utilities" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.234094 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="56db947e-810d-4bce-aefd-1d7e245299ff" containerName="registry-server" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.234103 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8745c7a-ec3a-4bc7-9ef1-7de335cf600a" containerName="registry-server" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.235632 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.245872 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6f568"] Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.384005 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-utilities\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.384577 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-catalog-content\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.384625 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k96tg\" (UniqueName: \"kubernetes.io/projected/c04b1b19-77b1-472d-83b8-37f0218ef691-kube-api-access-k96tg\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.486091 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k96tg\" (UniqueName: \"kubernetes.io/projected/c04b1b19-77b1-472d-83b8-37f0218ef691-kube-api-access-k96tg\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.486197 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-utilities\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.486323 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-catalog-content\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.486945 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-catalog-content\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.487099 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-utilities\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.515440 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k96tg\" (UniqueName: \"kubernetes.io/projected/c04b1b19-77b1-472d-83b8-37f0218ef691-kube-api-access-k96tg\") pod \"certified-operators-6f568\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:29 crc kubenswrapper[4799]: I0930 15:07:29.569532 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:30 crc kubenswrapper[4799]: I0930 15:07:30.178564 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6f568"] Sep 30 15:07:30 crc kubenswrapper[4799]: I0930 15:07:30.424263 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6f568" event={"ID":"c04b1b19-77b1-472d-83b8-37f0218ef691","Type":"ContainerStarted","Data":"ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9"} Sep 30 15:07:30 crc kubenswrapper[4799]: I0930 15:07:30.424631 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6f568" event={"ID":"c04b1b19-77b1-472d-83b8-37f0218ef691","Type":"ContainerStarted","Data":"011f43f7f62750747a1e146583231aa50377be010d979ed84b988489aabae07b"} Sep 30 15:07:31 crc kubenswrapper[4799]: I0930 15:07:31.436583 4799 generic.go:334] "Generic (PLEG): container finished" podID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerID="ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9" exitCode=0 Sep 30 15:07:31 crc kubenswrapper[4799]: I0930 15:07:31.436805 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6f568" event={"ID":"c04b1b19-77b1-472d-83b8-37f0218ef691","Type":"ContainerDied","Data":"ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9"} Sep 30 15:07:33 crc kubenswrapper[4799]: I0930 15:07:33.471238 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6f568" event={"ID":"c04b1b19-77b1-472d-83b8-37f0218ef691","Type":"ContainerStarted","Data":"ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8"} Sep 30 15:07:34 crc kubenswrapper[4799]: I0930 15:07:34.492234 4799 generic.go:334] "Generic (PLEG): container finished" podID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerID="ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8" exitCode=0 Sep 30 15:07:34 crc kubenswrapper[4799]: I0930 15:07:34.492458 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6f568" event={"ID":"c04b1b19-77b1-472d-83b8-37f0218ef691","Type":"ContainerDied","Data":"ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8"} Sep 30 15:07:35 crc kubenswrapper[4799]: I0930 15:07:35.525042 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6f568" event={"ID":"c04b1b19-77b1-472d-83b8-37f0218ef691","Type":"ContainerStarted","Data":"f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b"} Sep 30 15:07:39 crc kubenswrapper[4799]: I0930 15:07:39.571168 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:39 crc kubenswrapper[4799]: I0930 15:07:39.571783 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:39 crc kubenswrapper[4799]: I0930 15:07:39.621067 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:39 crc kubenswrapper[4799]: I0930 15:07:39.656968 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6f568" podStartSLOduration=7.060770337 podStartE2EDuration="10.656935078s" podCreationTimestamp="2025-09-30 15:07:29 +0000 UTC" firstStartedPulling="2025-09-30 15:07:31.441026445 +0000 UTC m=+2873.524626872" lastFinishedPulling="2025-09-30 15:07:35.037191186 +0000 UTC m=+2877.120791613" observedRunningTime="2025-09-30 15:07:35.549522923 +0000 UTC m=+2877.633123360" watchObservedRunningTime="2025-09-30 15:07:39.656935078 +0000 UTC m=+2881.740535505" Sep 30 15:07:40 crc kubenswrapper[4799]: I0930 15:07:40.653156 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:40 crc kubenswrapper[4799]: I0930 15:07:40.730389 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6f568"] Sep 30 15:07:42 crc kubenswrapper[4799]: I0930 15:07:42.608097 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6f568" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerName="registry-server" containerID="cri-o://f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b" gracePeriod=2 Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.107061 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.204167 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-utilities\") pod \"c04b1b19-77b1-472d-83b8-37f0218ef691\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.204366 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-catalog-content\") pod \"c04b1b19-77b1-472d-83b8-37f0218ef691\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.204463 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k96tg\" (UniqueName: \"kubernetes.io/projected/c04b1b19-77b1-472d-83b8-37f0218ef691-kube-api-access-k96tg\") pod \"c04b1b19-77b1-472d-83b8-37f0218ef691\" (UID: \"c04b1b19-77b1-472d-83b8-37f0218ef691\") " Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.205446 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-utilities" (OuterVolumeSpecName: "utilities") pod "c04b1b19-77b1-472d-83b8-37f0218ef691" (UID: "c04b1b19-77b1-472d-83b8-37f0218ef691"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.218369 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c04b1b19-77b1-472d-83b8-37f0218ef691-kube-api-access-k96tg" (OuterVolumeSpecName: "kube-api-access-k96tg") pod "c04b1b19-77b1-472d-83b8-37f0218ef691" (UID: "c04b1b19-77b1-472d-83b8-37f0218ef691"). InnerVolumeSpecName "kube-api-access-k96tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.268319 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c04b1b19-77b1-472d-83b8-37f0218ef691" (UID: "c04b1b19-77b1-472d-83b8-37f0218ef691"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.307104 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k96tg\" (UniqueName: \"kubernetes.io/projected/c04b1b19-77b1-472d-83b8-37f0218ef691-kube-api-access-k96tg\") on node \"crc\" DevicePath \"\"" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.307502 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.307590 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04b1b19-77b1-472d-83b8-37f0218ef691-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.620411 4799 generic.go:334] "Generic (PLEG): container finished" podID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerID="f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b" exitCode=0 Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.620473 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6f568" event={"ID":"c04b1b19-77b1-472d-83b8-37f0218ef691","Type":"ContainerDied","Data":"f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b"} Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.620518 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6f568" event={"ID":"c04b1b19-77b1-472d-83b8-37f0218ef691","Type":"ContainerDied","Data":"011f43f7f62750747a1e146583231aa50377be010d979ed84b988489aabae07b"} Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.620544 4799 scope.go:117] "RemoveContainer" containerID="f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.620564 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6f568" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.647824 4799 scope.go:117] "RemoveContainer" containerID="ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.692565 4799 scope.go:117] "RemoveContainer" containerID="ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.697707 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6f568"] Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.706436 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6f568"] Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.739720 4799 scope.go:117] "RemoveContainer" containerID="f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b" Sep 30 15:07:43 crc kubenswrapper[4799]: E0930 15:07:43.740579 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b\": container with ID starting with f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b not found: ID does not exist" containerID="f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.740755 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b"} err="failed to get container status \"f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b\": rpc error: code = NotFound desc = could not find container \"f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b\": container with ID starting with f9b9a1533e4f5982c331ab5ec30528f4ee72c492432e19dc5c30d4823ba8888b not found: ID does not exist" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.740877 4799 scope.go:117] "RemoveContainer" containerID="ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8" Sep 30 15:07:43 crc kubenswrapper[4799]: E0930 15:07:43.741475 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8\": container with ID starting with ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8 not found: ID does not exist" containerID="ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.741507 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8"} err="failed to get container status \"ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8\": rpc error: code = NotFound desc = could not find container \"ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8\": container with ID starting with ab19089d273d9565de5d5894f110ad97e5acf121146959d6bd2e627d477386b8 not found: ID does not exist" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.741526 4799 scope.go:117] "RemoveContainer" containerID="ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9" Sep 30 15:07:43 crc kubenswrapper[4799]: E0930 15:07:43.742159 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9\": container with ID starting with ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9 not found: ID does not exist" containerID="ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9" Sep 30 15:07:43 crc kubenswrapper[4799]: I0930 15:07:43.742204 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9"} err="failed to get container status \"ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9\": rpc error: code = NotFound desc = could not find container \"ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9\": container with ID starting with ac788f69f9f61cf73211519579964523ccec6cd783f9e51ec144a43a269a2ec9 not found: ID does not exist" Sep 30 15:07:44 crc kubenswrapper[4799]: I0930 15:07:44.524327 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" path="/var/lib/kubelet/pods/c04b1b19-77b1-472d-83b8-37f0218ef691/volumes" Sep 30 15:07:58 crc kubenswrapper[4799]: I0930 15:07:58.795589 4799 generic.go:334] "Generic (PLEG): container finished" podID="2da7cba0-9bbc-47c7-99b0-f204928886fa" containerID="3f4ee7a3ac8bbd5c6b28cecd6d03a66d61e719121e04c450f76ed0dd75c5d916" exitCode=0 Sep 30 15:07:58 crc kubenswrapper[4799]: I0930 15:07:58.795754 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" event={"ID":"2da7cba0-9bbc-47c7-99b0-f204928886fa","Type":"ContainerDied","Data":"3f4ee7a3ac8bbd5c6b28cecd6d03a66d61e719121e04c450f76ed0dd75c5d916"} Sep 30 15:07:59 crc kubenswrapper[4799]: I0930 15:07:59.648707 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:07:59 crc kubenswrapper[4799]: I0930 15:07:59.648791 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.325415 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.501811 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-extra-config-0\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.501938 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjgmm\" (UniqueName: \"kubernetes.io/projected/2da7cba0-9bbc-47c7-99b0-f204928886fa-kube-api-access-qjgmm\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.501973 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-1\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.502867 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-0\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.503115 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-1\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.503185 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-0\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.503231 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-inventory\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.503284 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-combined-ca-bundle\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.503334 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-ssh-key\") pod \"2da7cba0-9bbc-47c7-99b0-f204928886fa\" (UID: \"2da7cba0-9bbc-47c7-99b0-f204928886fa\") " Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.508877 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da7cba0-9bbc-47c7-99b0-f204928886fa-kube-api-access-qjgmm" (OuterVolumeSpecName: "kube-api-access-qjgmm") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "kube-api-access-qjgmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.530610 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.544636 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.556231 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.560532 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.561480 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.561725 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-inventory" (OuterVolumeSpecName: "inventory") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.576910 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.578905 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2da7cba0-9bbc-47c7-99b0-f204928886fa" (UID: "2da7cba0-9bbc-47c7-99b0-f204928886fa"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.605868 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjgmm\" (UniqueName: \"kubernetes.io/projected/2da7cba0-9bbc-47c7-99b0-f204928886fa-kube-api-access-qjgmm\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.605915 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.605932 4799 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.605950 4799 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.605964 4799 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.605978 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.605991 4799 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.606002 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2da7cba0-9bbc-47c7-99b0-f204928886fa-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.606016 4799 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2da7cba0-9bbc-47c7-99b0-f204928886fa-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.819081 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" event={"ID":"2da7cba0-9bbc-47c7-99b0-f204928886fa","Type":"ContainerDied","Data":"b2d1f15412cf4c7767feccea56c10fd438e7e2fa432e6a0b194f0b79db1bed41"} Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.819473 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2d1f15412cf4c7767feccea56c10fd438e7e2fa432e6a0b194f0b79db1bed41" Sep 30 15:08:00 crc kubenswrapper[4799]: I0930 15:08:00.819136 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tlstx" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.033056 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr"] Sep 30 15:08:01 crc kubenswrapper[4799]: E0930 15:08:01.033601 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerName="extract-utilities" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.033631 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerName="extract-utilities" Sep 30 15:08:01 crc kubenswrapper[4799]: E0930 15:08:01.033686 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerName="registry-server" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.033696 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerName="registry-server" Sep 30 15:08:01 crc kubenswrapper[4799]: E0930 15:08:01.033708 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerName="extract-content" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.033716 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerName="extract-content" Sep 30 15:08:01 crc kubenswrapper[4799]: E0930 15:08:01.033734 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da7cba0-9bbc-47c7-99b0-f204928886fa" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.033742 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da7cba0-9bbc-47c7-99b0-f204928886fa" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.033966 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da7cba0-9bbc-47c7-99b0-f204928886fa" containerName="nova-edpm-deployment-openstack-edpm-ipam" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.034001 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c04b1b19-77b1-472d-83b8-37f0218ef691" containerName="registry-server" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.034859 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.037398 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xqfzw" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.037551 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.037606 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.038637 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.040295 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.059468 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr"] Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.217024 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.217230 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.217502 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.217598 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcdbk\" (UniqueName: \"kubernetes.io/projected/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-kube-api-access-fcdbk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.217717 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.217759 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.217790 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.320019 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.320094 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.320116 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.320137 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.320181 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.320287 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.320317 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcdbk\" (UniqueName: \"kubernetes.io/projected/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-kube-api-access-fcdbk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.325595 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.325669 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.326802 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.326805 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.327760 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.332171 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.342699 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcdbk\" (UniqueName: \"kubernetes.io/projected/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-kube-api-access-fcdbk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-f72vr\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.352336 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:08:01 crc kubenswrapper[4799]: I0930 15:08:01.929693 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr"] Sep 30 15:08:02 crc kubenswrapper[4799]: I0930 15:08:02.838640 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" event={"ID":"a86a2d67-7c0c-44b5-a435-15fccfd80d5e","Type":"ContainerStarted","Data":"db2b63b299e13d39f5c7f40ebb24ef0f06a8fbbbdd6949d09a7b03dd4c9efd69"} Sep 30 15:08:02 crc kubenswrapper[4799]: I0930 15:08:02.839841 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" event={"ID":"a86a2d67-7c0c-44b5-a435-15fccfd80d5e","Type":"ContainerStarted","Data":"f813e86af5ae4401b232f91272e9e389d905b3d059b168327b85334f4f79d67e"} Sep 30 15:08:02 crc kubenswrapper[4799]: I0930 15:08:02.868188 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" podStartSLOduration=1.6895647070000002 podStartE2EDuration="1.868161879s" podCreationTimestamp="2025-09-30 15:08:01 +0000 UTC" firstStartedPulling="2025-09-30 15:08:01.938827812 +0000 UTC m=+2904.022428239" lastFinishedPulling="2025-09-30 15:08:02.117424994 +0000 UTC m=+2904.201025411" observedRunningTime="2025-09-30 15:08:02.860727181 +0000 UTC m=+2904.944327618" watchObservedRunningTime="2025-09-30 15:08:02.868161879 +0000 UTC m=+2904.951762316" Sep 30 15:08:29 crc kubenswrapper[4799]: I0930 15:08:29.648770 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:08:29 crc kubenswrapper[4799]: I0930 15:08:29.649395 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:08:59 crc kubenswrapper[4799]: I0930 15:08:59.648489 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:08:59 crc kubenswrapper[4799]: I0930 15:08:59.649143 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:08:59 crc kubenswrapper[4799]: I0930 15:08:59.649213 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:08:59 crc kubenswrapper[4799]: I0930 15:08:59.650219 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:08:59 crc kubenswrapper[4799]: I0930 15:08:59.650294 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" gracePeriod=600 Sep 30 15:08:59 crc kubenswrapper[4799]: E0930 15:08:59.784489 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:09:00 crc kubenswrapper[4799]: I0930 15:09:00.473785 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" exitCode=0 Sep 30 15:09:00 crc kubenswrapper[4799]: I0930 15:09:00.475041 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53"} Sep 30 15:09:00 crc kubenswrapper[4799]: I0930 15:09:00.475132 4799 scope.go:117] "RemoveContainer" containerID="8ab6660d9eab2d41fdd2de146c8c871d31c86110d25066571e7ae3aa2faee410" Sep 30 15:09:00 crc kubenswrapper[4799]: I0930 15:09:00.476967 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:09:00 crc kubenswrapper[4799]: E0930 15:09:00.477776 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:09:13 crc kubenswrapper[4799]: I0930 15:09:13.503692 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:09:13 crc kubenswrapper[4799]: E0930 15:09:13.504566 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:09:24 crc kubenswrapper[4799]: I0930 15:09:24.504054 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:09:24 crc kubenswrapper[4799]: E0930 15:09:24.505188 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:09:37 crc kubenswrapper[4799]: I0930 15:09:37.504712 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:09:37 crc kubenswrapper[4799]: E0930 15:09:37.505733 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:09:50 crc kubenswrapper[4799]: I0930 15:09:50.503234 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:09:50 crc kubenswrapper[4799]: E0930 15:09:50.504111 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:10:04 crc kubenswrapper[4799]: I0930 15:10:04.505862 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:10:04 crc kubenswrapper[4799]: E0930 15:10:04.506836 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:10:16 crc kubenswrapper[4799]: I0930 15:10:16.504165 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:10:16 crc kubenswrapper[4799]: E0930 15:10:16.505105 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:10:30 crc kubenswrapper[4799]: I0930 15:10:30.503893 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:10:30 crc kubenswrapper[4799]: E0930 15:10:30.505778 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:10:43 crc kubenswrapper[4799]: I0930 15:10:43.503851 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:10:43 crc kubenswrapper[4799]: E0930 15:10:43.506608 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:10:57 crc kubenswrapper[4799]: I0930 15:10:57.504728 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:10:57 crc kubenswrapper[4799]: E0930 15:10:57.506161 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:11:12 crc kubenswrapper[4799]: I0930 15:11:12.505591 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:11:12 crc kubenswrapper[4799]: E0930 15:11:12.507563 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:11:24 crc kubenswrapper[4799]: I0930 15:11:24.504589 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:11:24 crc kubenswrapper[4799]: E0930 15:11:24.506274 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:11:27 crc kubenswrapper[4799]: I0930 15:11:27.030036 4799 generic.go:334] "Generic (PLEG): container finished" podID="a86a2d67-7c0c-44b5-a435-15fccfd80d5e" containerID="db2b63b299e13d39f5c7f40ebb24ef0f06a8fbbbdd6949d09a7b03dd4c9efd69" exitCode=0 Sep 30 15:11:27 crc kubenswrapper[4799]: I0930 15:11:27.030131 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" event={"ID":"a86a2d67-7c0c-44b5-a435-15fccfd80d5e","Type":"ContainerDied","Data":"db2b63b299e13d39f5c7f40ebb24ef0f06a8fbbbdd6949d09a7b03dd4c9efd69"} Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.541595 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.708487 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcdbk\" (UniqueName: \"kubernetes.io/projected/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-kube-api-access-fcdbk\") pod \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.708732 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ssh-key\") pod \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.708766 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-1\") pod \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.708838 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-2\") pod \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.708884 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-telemetry-combined-ca-bundle\") pod \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.709902 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-inventory\") pod \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.710073 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-0\") pod \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\" (UID: \"a86a2d67-7c0c-44b5-a435-15fccfd80d5e\") " Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.720835 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a86a2d67-7c0c-44b5-a435-15fccfd80d5e" (UID: "a86a2d67-7c0c-44b5-a435-15fccfd80d5e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.723626 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-kube-api-access-fcdbk" (OuterVolumeSpecName: "kube-api-access-fcdbk") pod "a86a2d67-7c0c-44b5-a435-15fccfd80d5e" (UID: "a86a2d67-7c0c-44b5-a435-15fccfd80d5e"). InnerVolumeSpecName "kube-api-access-fcdbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.746113 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-inventory" (OuterVolumeSpecName: "inventory") pod "a86a2d67-7c0c-44b5-a435-15fccfd80d5e" (UID: "a86a2d67-7c0c-44b5-a435-15fccfd80d5e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.747113 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a86a2d67-7c0c-44b5-a435-15fccfd80d5e" (UID: "a86a2d67-7c0c-44b5-a435-15fccfd80d5e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.748184 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "a86a2d67-7c0c-44b5-a435-15fccfd80d5e" (UID: "a86a2d67-7c0c-44b5-a435-15fccfd80d5e"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.761261 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "a86a2d67-7c0c-44b5-a435-15fccfd80d5e" (UID: "a86a2d67-7c0c-44b5-a435-15fccfd80d5e"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.762123 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "a86a2d67-7c0c-44b5-a435-15fccfd80d5e" (UID: "a86a2d67-7c0c-44b5-a435-15fccfd80d5e"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.813936 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcdbk\" (UniqueName: \"kubernetes.io/projected/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-kube-api-access-fcdbk\") on node \"crc\" DevicePath \"\"" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.813990 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.814007 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.814020 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.814030 4799 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.814043 4799 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-inventory\") on node \"crc\" DevicePath \"\"" Sep 30 15:11:28 crc kubenswrapper[4799]: I0930 15:11:28.814053 4799 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a86a2d67-7c0c-44b5-a435-15fccfd80d5e-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Sep 30 15:11:29 crc kubenswrapper[4799]: I0930 15:11:29.057724 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" event={"ID":"a86a2d67-7c0c-44b5-a435-15fccfd80d5e","Type":"ContainerDied","Data":"f813e86af5ae4401b232f91272e9e389d905b3d059b168327b85334f4f79d67e"} Sep 30 15:11:29 crc kubenswrapper[4799]: I0930 15:11:29.057772 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f813e86af5ae4401b232f91272e9e389d905b3d059b168327b85334f4f79d67e" Sep 30 15:11:29 crc kubenswrapper[4799]: I0930 15:11:29.057868 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-f72vr" Sep 30 15:11:39 crc kubenswrapper[4799]: I0930 15:11:39.503947 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:11:39 crc kubenswrapper[4799]: E0930 15:11:39.505297 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:11:54 crc kubenswrapper[4799]: I0930 15:11:54.521036 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:11:54 crc kubenswrapper[4799]: E0930 15:11:54.522895 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:12:08 crc kubenswrapper[4799]: I0930 15:12:08.513457 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:12:08 crc kubenswrapper[4799]: E0930 15:12:08.516998 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:12:19 crc kubenswrapper[4799]: I0930 15:12:19.503780 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:12:19 crc kubenswrapper[4799]: E0930 15:12:19.505153 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.502095 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 15:12:30 crc kubenswrapper[4799]: E0930 15:12:30.507604 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86a2d67-7c0c-44b5-a435-15fccfd80d5e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.507631 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86a2d67-7c0c-44b5-a435-15fccfd80d5e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.507904 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86a2d67-7c0c-44b5-a435-15fccfd80d5e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.507971 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:12:30 crc kubenswrapper[4799]: E0930 15:12:30.508275 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.508902 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.514016 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5f8xr" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.515765 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.515829 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.517177 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.555504 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640031 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640119 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640169 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640247 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640277 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640383 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640447 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz4m8\" (UniqueName: \"kubernetes.io/projected/ab210e62-906f-46ad-b80a-fbf00a875124-kube-api-access-nz4m8\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640476 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-config-data\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.640518 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.742903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz4m8\" (UniqueName: \"kubernetes.io/projected/ab210e62-906f-46ad-b80a-fbf00a875124-kube-api-access-nz4m8\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.742977 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-config-data\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.743008 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.743145 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.743190 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.743246 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.743373 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.743410 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.743495 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.744832 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-config-data\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.744924 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.744951 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.745263 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.745793 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.751890 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.755619 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.757440 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.769369 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz4m8\" (UniqueName: \"kubernetes.io/projected/ab210e62-906f-46ad-b80a-fbf00a875124-kube-api-access-nz4m8\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.784604 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " pod="openstack/tempest-tests-tempest" Sep 30 15:12:30 crc kubenswrapper[4799]: I0930 15:12:30.843743 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 15:12:31 crc kubenswrapper[4799]: I0930 15:12:31.356005 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Sep 30 15:12:31 crc kubenswrapper[4799]: I0930 15:12:31.366620 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 15:12:31 crc kubenswrapper[4799]: I0930 15:12:31.750095 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ab210e62-906f-46ad-b80a-fbf00a875124","Type":"ContainerStarted","Data":"a8a4f366047149510b6bd1463e01c76a42804713ad28e9300ce24a04b64f8d89"} Sep 30 15:12:43 crc kubenswrapper[4799]: I0930 15:12:43.506368 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:12:43 crc kubenswrapper[4799]: E0930 15:12:43.507249 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:12:55 crc kubenswrapper[4799]: I0930 15:12:55.505724 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:12:55 crc kubenswrapper[4799]: E0930 15:12:55.506976 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:13:09 crc kubenswrapper[4799]: I0930 15:13:09.504346 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:13:09 crc kubenswrapper[4799]: E0930 15:13:09.505241 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:13:20 crc kubenswrapper[4799]: E0930 15:13:20.523122 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Sep 30 15:13:20 crc kubenswrapper[4799]: E0930 15:13:20.527839 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nz4m8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(ab210e62-906f-46ad-b80a-fbf00a875124): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Sep 30 15:13:20 crc kubenswrapper[4799]: E0930 15:13:20.528985 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="ab210e62-906f-46ad-b80a-fbf00a875124" Sep 30 15:13:21 crc kubenswrapper[4799]: E0930 15:13:21.337296 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="ab210e62-906f-46ad-b80a-fbf00a875124" Sep 30 15:13:24 crc kubenswrapper[4799]: I0930 15:13:24.503641 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:13:24 crc kubenswrapper[4799]: E0930 15:13:24.504364 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:13:33 crc kubenswrapper[4799]: I0930 15:13:33.095996 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Sep 30 15:13:34 crc kubenswrapper[4799]: I0930 15:13:34.519212 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ab210e62-906f-46ad-b80a-fbf00a875124","Type":"ContainerStarted","Data":"b8c2ed383ec607aa5c232aa30bcce5d5ad65e2ca9a50987192da6ae49739f4ab"} Sep 30 15:13:34 crc kubenswrapper[4799]: I0930 15:13:34.549510 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.823460684 podStartE2EDuration="1m5.549479378s" podCreationTimestamp="2025-09-30 15:12:29 +0000 UTC" firstStartedPulling="2025-09-30 15:12:31.366255815 +0000 UTC m=+3173.449856252" lastFinishedPulling="2025-09-30 15:13:33.092274519 +0000 UTC m=+3235.175874946" observedRunningTime="2025-09-30 15:13:34.541913786 +0000 UTC m=+3236.625514233" watchObservedRunningTime="2025-09-30 15:13:34.549479378 +0000 UTC m=+3236.633079805" Sep 30 15:13:37 crc kubenswrapper[4799]: I0930 15:13:37.503301 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:13:37 crc kubenswrapper[4799]: E0930 15:13:37.503915 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:13:49 crc kubenswrapper[4799]: I0930 15:13:49.504990 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:13:49 crc kubenswrapper[4799]: E0930 15:13:49.505753 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:14:02 crc kubenswrapper[4799]: I0930 15:14:02.508206 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:14:02 crc kubenswrapper[4799]: I0930 15:14:02.850387 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"12492a0f5c49ea303b04b6e5eff4f9f2b1a2ef113c159201aca5e1132a65a941"} Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.192237 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm"] Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.195197 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.198390 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.198860 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.296889 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm"] Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.324629 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-secret-volume\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.326148 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8xqp\" (UniqueName: \"kubernetes.io/projected/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-kube-api-access-g8xqp\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.326377 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-config-volume\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.428474 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8xqp\" (UniqueName: \"kubernetes.io/projected/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-kube-api-access-g8xqp\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.428600 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-config-volume\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.429721 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-secret-volume\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.431521 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-config-volume\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.451030 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-secret-volume\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.455252 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8xqp\" (UniqueName: \"kubernetes.io/projected/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-kube-api-access-g8xqp\") pod \"collect-profiles-29320755-t54tm\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:00 crc kubenswrapper[4799]: I0930 15:15:00.523686 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:02 crc kubenswrapper[4799]: I0930 15:15:02.405060 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm"] Sep 30 15:15:02 crc kubenswrapper[4799]: I0930 15:15:02.646710 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" event={"ID":"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e","Type":"ContainerStarted","Data":"2054439a6eae71397a5af512879e8b42619dc63bc649fb3887a5e6498bcd8c92"} Sep 30 15:15:03 crc kubenswrapper[4799]: I0930 15:15:03.661415 4799 generic.go:334] "Generic (PLEG): container finished" podID="01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e" containerID="bfdc28f364632f12d801f1b97a24863f624681df301bd32b39cd238fea49ac89" exitCode=0 Sep 30 15:15:03 crc kubenswrapper[4799]: I0930 15:15:03.661498 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" event={"ID":"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e","Type":"ContainerDied","Data":"bfdc28f364632f12d801f1b97a24863f624681df301bd32b39cd238fea49ac89"} Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.211797 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.371427 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-secret-volume\") pod \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.371807 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8xqp\" (UniqueName: \"kubernetes.io/projected/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-kube-api-access-g8xqp\") pod \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.371877 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-config-volume\") pod \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\" (UID: \"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e\") " Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.373106 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-config-volume" (OuterVolumeSpecName: "config-volume") pod "01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e" (UID: "01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.383501 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-kube-api-access-g8xqp" (OuterVolumeSpecName: "kube-api-access-g8xqp") pod "01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e" (UID: "01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e"). InnerVolumeSpecName "kube-api-access-g8xqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.393988 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e" (UID: "01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.475514 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8xqp\" (UniqueName: \"kubernetes.io/projected/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-kube-api-access-g8xqp\") on node \"crc\" DevicePath \"\"" Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.475576 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.475591 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.688819 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" event={"ID":"01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e","Type":"ContainerDied","Data":"2054439a6eae71397a5af512879e8b42619dc63bc649fb3887a5e6498bcd8c92"} Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.689317 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2054439a6eae71397a5af512879e8b42619dc63bc649fb3887a5e6498bcd8c92" Sep 30 15:15:05 crc kubenswrapper[4799]: I0930 15:15:05.689421 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320755-t54tm" Sep 30 15:15:06 crc kubenswrapper[4799]: I0930 15:15:06.327514 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd"] Sep 30 15:15:06 crc kubenswrapper[4799]: I0930 15:15:06.337458 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320710-9r4wd"] Sep 30 15:15:06 crc kubenswrapper[4799]: I0930 15:15:06.517219 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5508bbb-b87a-4e2e-bc55-940edda975cc" path="/var/lib/kubelet/pods/a5508bbb-b87a-4e2e-bc55-940edda975cc/volumes" Sep 30 15:15:12 crc kubenswrapper[4799]: I0930 15:15:12.612810 4799 scope.go:117] "RemoveContainer" containerID="414693d8ed85d684c76c8a5bd0bc96d3fd2d1dbb8df38e5f1ad928d3dd2993fa" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.622304 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m985h"] Sep 30 15:15:37 crc kubenswrapper[4799]: E0930 15:15:37.623295 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e" containerName="collect-profiles" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.623311 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e" containerName="collect-profiles" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.623510 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="01426b3b-9eb3-4e08-8c2d-ca8a6b1cee0e" containerName="collect-profiles" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.625064 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.638846 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m985h"] Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.736984 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-utilities\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.737381 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-catalog-content\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.737464 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b7g5\" (UniqueName: \"kubernetes.io/projected/7355e5ad-8d94-464f-b306-5b3410fe9779-kube-api-access-8b7g5\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.840436 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-catalog-content\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.840552 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b7g5\" (UniqueName: \"kubernetes.io/projected/7355e5ad-8d94-464f-b306-5b3410fe9779-kube-api-access-8b7g5\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.841120 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-catalog-content\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.841242 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-utilities\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.841625 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-utilities\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.875348 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b7g5\" (UniqueName: \"kubernetes.io/projected/7355e5ad-8d94-464f-b306-5b3410fe9779-kube-api-access-8b7g5\") pod \"redhat-operators-m985h\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:37 crc kubenswrapper[4799]: I0930 15:15:37.966228 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:38 crc kubenswrapper[4799]: I0930 15:15:38.612383 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m985h"] Sep 30 15:15:39 crc kubenswrapper[4799]: I0930 15:15:39.103487 4799 generic.go:334] "Generic (PLEG): container finished" podID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerID="414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6" exitCode=0 Sep 30 15:15:39 crc kubenswrapper[4799]: I0930 15:15:39.103601 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m985h" event={"ID":"7355e5ad-8d94-464f-b306-5b3410fe9779","Type":"ContainerDied","Data":"414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6"} Sep 30 15:15:39 crc kubenswrapper[4799]: I0930 15:15:39.104069 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m985h" event={"ID":"7355e5ad-8d94-464f-b306-5b3410fe9779","Type":"ContainerStarted","Data":"a5f249a1600a4147966c9c43e3ab3c8ecefd156a9e019d931c9f551e4565ffb4"} Sep 30 15:15:41 crc kubenswrapper[4799]: I0930 15:15:41.130342 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m985h" event={"ID":"7355e5ad-8d94-464f-b306-5b3410fe9779","Type":"ContainerStarted","Data":"ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e"} Sep 30 15:15:46 crc kubenswrapper[4799]: I0930 15:15:46.195924 4799 generic.go:334] "Generic (PLEG): container finished" podID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerID="ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e" exitCode=0 Sep 30 15:15:46 crc kubenswrapper[4799]: I0930 15:15:46.196517 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m985h" event={"ID":"7355e5ad-8d94-464f-b306-5b3410fe9779","Type":"ContainerDied","Data":"ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e"} Sep 30 15:15:47 crc kubenswrapper[4799]: I0930 15:15:47.208256 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m985h" event={"ID":"7355e5ad-8d94-464f-b306-5b3410fe9779","Type":"ContainerStarted","Data":"8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8"} Sep 30 15:15:47 crc kubenswrapper[4799]: I0930 15:15:47.231386 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m985h" podStartSLOduration=2.668833718 podStartE2EDuration="10.231363676s" podCreationTimestamp="2025-09-30 15:15:37 +0000 UTC" firstStartedPulling="2025-09-30 15:15:39.106232845 +0000 UTC m=+3361.189833272" lastFinishedPulling="2025-09-30 15:15:46.668762813 +0000 UTC m=+3368.752363230" observedRunningTime="2025-09-30 15:15:47.227177214 +0000 UTC m=+3369.310777641" watchObservedRunningTime="2025-09-30 15:15:47.231363676 +0000 UTC m=+3369.314964103" Sep 30 15:15:47 crc kubenswrapper[4799]: I0930 15:15:47.966446 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:47 crc kubenswrapper[4799]: I0930 15:15:47.967777 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:15:49 crc kubenswrapper[4799]: I0930 15:15:49.023818 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m985h" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="registry-server" probeResult="failure" output=< Sep 30 15:15:49 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:15:49 crc kubenswrapper[4799]: > Sep 30 15:15:59 crc kubenswrapper[4799]: I0930 15:15:59.032579 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m985h" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="registry-server" probeResult="failure" output=< Sep 30 15:15:59 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:15:59 crc kubenswrapper[4799]: > Sep 30 15:16:09 crc kubenswrapper[4799]: I0930 15:16:09.036918 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m985h" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="registry-server" probeResult="failure" output=< Sep 30 15:16:09 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:16:09 crc kubenswrapper[4799]: > Sep 30 15:16:18 crc kubenswrapper[4799]: I0930 15:16:18.043311 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:16:18 crc kubenswrapper[4799]: I0930 15:16:18.108362 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:16:18 crc kubenswrapper[4799]: I0930 15:16:18.302773 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m985h"] Sep 30 15:16:19 crc kubenswrapper[4799]: I0930 15:16:19.558759 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m985h" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="registry-server" containerID="cri-o://8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8" gracePeriod=2 Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.426193 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.487291 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-utilities\") pod \"7355e5ad-8d94-464f-b306-5b3410fe9779\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.487365 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b7g5\" (UniqueName: \"kubernetes.io/projected/7355e5ad-8d94-464f-b306-5b3410fe9779-kube-api-access-8b7g5\") pod \"7355e5ad-8d94-464f-b306-5b3410fe9779\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.487855 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-catalog-content\") pod \"7355e5ad-8d94-464f-b306-5b3410fe9779\" (UID: \"7355e5ad-8d94-464f-b306-5b3410fe9779\") " Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.494117 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-utilities" (OuterVolumeSpecName: "utilities") pod "7355e5ad-8d94-464f-b306-5b3410fe9779" (UID: "7355e5ad-8d94-464f-b306-5b3410fe9779"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.512957 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7355e5ad-8d94-464f-b306-5b3410fe9779-kube-api-access-8b7g5" (OuterVolumeSpecName: "kube-api-access-8b7g5") pod "7355e5ad-8d94-464f-b306-5b3410fe9779" (UID: "7355e5ad-8d94-464f-b306-5b3410fe9779"). InnerVolumeSpecName "kube-api-access-8b7g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.592387 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.594102 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b7g5\" (UniqueName: \"kubernetes.io/projected/7355e5ad-8d94-464f-b306-5b3410fe9779-kube-api-access-8b7g5\") on node \"crc\" DevicePath \"\"" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.603245 4799 generic.go:334] "Generic (PLEG): container finished" podID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerID="8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8" exitCode=0 Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.603322 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m985h" event={"ID":"7355e5ad-8d94-464f-b306-5b3410fe9779","Type":"ContainerDied","Data":"8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8"} Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.603379 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m985h" event={"ID":"7355e5ad-8d94-464f-b306-5b3410fe9779","Type":"ContainerDied","Data":"a5f249a1600a4147966c9c43e3ab3c8ecefd156a9e019d931c9f551e4565ffb4"} Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.603409 4799 scope.go:117] "RemoveContainer" containerID="8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.603713 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m985h" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.658126 4799 scope.go:117] "RemoveContainer" containerID="ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.669403 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7355e5ad-8d94-464f-b306-5b3410fe9779" (UID: "7355e5ad-8d94-464f-b306-5b3410fe9779"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.696069 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7355e5ad-8d94-464f-b306-5b3410fe9779-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.711138 4799 scope.go:117] "RemoveContainer" containerID="414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.768895 4799 scope.go:117] "RemoveContainer" containerID="8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8" Sep 30 15:16:20 crc kubenswrapper[4799]: E0930 15:16:20.770520 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8\": container with ID starting with 8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8 not found: ID does not exist" containerID="8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.770550 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8"} err="failed to get container status \"8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8\": rpc error: code = NotFound desc = could not find container \"8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8\": container with ID starting with 8615829adada3c4f59026dcc2035f98dbf3b6450434997ba02cc06aa19d48ff8 not found: ID does not exist" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.770578 4799 scope.go:117] "RemoveContainer" containerID="ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e" Sep 30 15:16:20 crc kubenswrapper[4799]: E0930 15:16:20.770991 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e\": container with ID starting with ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e not found: ID does not exist" containerID="ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.771015 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e"} err="failed to get container status \"ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e\": rpc error: code = NotFound desc = could not find container \"ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e\": container with ID starting with ab070ba1c7c56a35d132e6e8a9e671a1cd53bcc406aedef24fc4f5a01ce86b1e not found: ID does not exist" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.771032 4799 scope.go:117] "RemoveContainer" containerID="414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6" Sep 30 15:16:20 crc kubenswrapper[4799]: E0930 15:16:20.771344 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6\": container with ID starting with 414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6 not found: ID does not exist" containerID="414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.771365 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6"} err="failed to get container status \"414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6\": rpc error: code = NotFound desc = could not find container \"414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6\": container with ID starting with 414570579f70675e3344f200b342fddc610a14b43d2cc1b7af16671677d878b6 not found: ID does not exist" Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.940174 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m985h"] Sep 30 15:16:20 crc kubenswrapper[4799]: I0930 15:16:20.956299 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m985h"] Sep 30 15:16:22 crc kubenswrapper[4799]: I0930 15:16:22.518197 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" path="/var/lib/kubelet/pods/7355e5ad-8d94-464f-b306-5b3410fe9779/volumes" Sep 30 15:16:29 crc kubenswrapper[4799]: I0930 15:16:29.648410 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:16:29 crc kubenswrapper[4799]: I0930 15:16:29.648957 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.280470 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tmdr9"] Sep 30 15:16:53 crc kubenswrapper[4799]: E0930 15:16:53.281624 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="registry-server" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.281645 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="registry-server" Sep 30 15:16:53 crc kubenswrapper[4799]: E0930 15:16:53.281679 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="extract-content" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.281687 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="extract-content" Sep 30 15:16:53 crc kubenswrapper[4799]: E0930 15:16:53.281720 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="extract-utilities" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.281747 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="extract-utilities" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.281966 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7355e5ad-8d94-464f-b306-5b3410fe9779" containerName="registry-server" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.283881 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.301293 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmdr9"] Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.470661 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv9gr\" (UniqueName: \"kubernetes.io/projected/9df2fce8-7ad6-4572-8970-bcf915abc641-kube-api-access-nv9gr\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.470741 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-catalog-content\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.471013 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-utilities\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.572850 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-utilities\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.573015 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv9gr\" (UniqueName: \"kubernetes.io/projected/9df2fce8-7ad6-4572-8970-bcf915abc641-kube-api-access-nv9gr\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.573067 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-catalog-content\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.573821 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-catalog-content\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.573990 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-utilities\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.597641 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv9gr\" (UniqueName: \"kubernetes.io/projected/9df2fce8-7ad6-4572-8970-bcf915abc641-kube-api-access-nv9gr\") pod \"redhat-marketplace-tmdr9\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:53 crc kubenswrapper[4799]: I0930 15:16:53.609390 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:16:54 crc kubenswrapper[4799]: I0930 15:16:54.201029 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmdr9"] Sep 30 15:16:54 crc kubenswrapper[4799]: I0930 15:16:54.970576 4799 generic.go:334] "Generic (PLEG): container finished" podID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerID="48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2" exitCode=0 Sep 30 15:16:54 crc kubenswrapper[4799]: I0930 15:16:54.972421 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmdr9" event={"ID":"9df2fce8-7ad6-4572-8970-bcf915abc641","Type":"ContainerDied","Data":"48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2"} Sep 30 15:16:54 crc kubenswrapper[4799]: I0930 15:16:54.972454 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmdr9" event={"ID":"9df2fce8-7ad6-4572-8970-bcf915abc641","Type":"ContainerStarted","Data":"d63d89dd8d38a883a81f2f423c8a532f563fd4ab6e68eb58db913ac3b570bb57"} Sep 30 15:16:55 crc kubenswrapper[4799]: I0930 15:16:55.982932 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmdr9" event={"ID":"9df2fce8-7ad6-4572-8970-bcf915abc641","Type":"ContainerStarted","Data":"d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e"} Sep 30 15:16:56 crc kubenswrapper[4799]: I0930 15:16:56.997072 4799 generic.go:334] "Generic (PLEG): container finished" podID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerID="d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e" exitCode=0 Sep 30 15:16:56 crc kubenswrapper[4799]: I0930 15:16:56.997450 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmdr9" event={"ID":"9df2fce8-7ad6-4572-8970-bcf915abc641","Type":"ContainerDied","Data":"d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e"} Sep 30 15:16:58 crc kubenswrapper[4799]: I0930 15:16:58.013180 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmdr9" event={"ID":"9df2fce8-7ad6-4572-8970-bcf915abc641","Type":"ContainerStarted","Data":"3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427"} Sep 30 15:16:59 crc kubenswrapper[4799]: I0930 15:16:59.648549 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:16:59 crc kubenswrapper[4799]: I0930 15:16:59.648945 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:17:03 crc kubenswrapper[4799]: I0930 15:17:03.610405 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:17:03 crc kubenswrapper[4799]: I0930 15:17:03.612673 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:17:03 crc kubenswrapper[4799]: I0930 15:17:03.674816 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:17:03 crc kubenswrapper[4799]: I0930 15:17:03.699553 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tmdr9" podStartSLOduration=8.185039684 podStartE2EDuration="10.699518065s" podCreationTimestamp="2025-09-30 15:16:53 +0000 UTC" firstStartedPulling="2025-09-30 15:16:54.97300544 +0000 UTC m=+3437.056605867" lastFinishedPulling="2025-09-30 15:16:57.487483821 +0000 UTC m=+3439.571084248" observedRunningTime="2025-09-30 15:16:58.048150403 +0000 UTC m=+3440.131750950" watchObservedRunningTime="2025-09-30 15:17:03.699518065 +0000 UTC m=+3445.783118492" Sep 30 15:17:04 crc kubenswrapper[4799]: I0930 15:17:04.145129 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:17:04 crc kubenswrapper[4799]: I0930 15:17:04.263945 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmdr9"] Sep 30 15:17:06 crc kubenswrapper[4799]: I0930 15:17:06.097303 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tmdr9" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerName="registry-server" containerID="cri-o://3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427" gracePeriod=2 Sep 30 15:17:06 crc kubenswrapper[4799]: I0930 15:17:06.888499 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:17:06 crc kubenswrapper[4799]: I0930 15:17:06.991463 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv9gr\" (UniqueName: \"kubernetes.io/projected/9df2fce8-7ad6-4572-8970-bcf915abc641-kube-api-access-nv9gr\") pod \"9df2fce8-7ad6-4572-8970-bcf915abc641\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " Sep 30 15:17:06 crc kubenswrapper[4799]: I0930 15:17:06.991564 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-utilities\") pod \"9df2fce8-7ad6-4572-8970-bcf915abc641\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " Sep 30 15:17:06 crc kubenswrapper[4799]: I0930 15:17:06.991606 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-catalog-content\") pod \"9df2fce8-7ad6-4572-8970-bcf915abc641\" (UID: \"9df2fce8-7ad6-4572-8970-bcf915abc641\") " Sep 30 15:17:06 crc kubenswrapper[4799]: I0930 15:17:06.993020 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-utilities" (OuterVolumeSpecName: "utilities") pod "9df2fce8-7ad6-4572-8970-bcf915abc641" (UID: "9df2fce8-7ad6-4572-8970-bcf915abc641"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.006606 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9df2fce8-7ad6-4572-8970-bcf915abc641" (UID: "9df2fce8-7ad6-4572-8970-bcf915abc641"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.008231 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9df2fce8-7ad6-4572-8970-bcf915abc641-kube-api-access-nv9gr" (OuterVolumeSpecName: "kube-api-access-nv9gr") pod "9df2fce8-7ad6-4572-8970-bcf915abc641" (UID: "9df2fce8-7ad6-4572-8970-bcf915abc641"). InnerVolumeSpecName "kube-api-access-nv9gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.094476 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv9gr\" (UniqueName: \"kubernetes.io/projected/9df2fce8-7ad6-4572-8970-bcf915abc641-kube-api-access-nv9gr\") on node \"crc\" DevicePath \"\"" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.094539 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.094550 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df2fce8-7ad6-4572-8970-bcf915abc641-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.113783 4799 generic.go:334] "Generic (PLEG): container finished" podID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerID="3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427" exitCode=0 Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.113873 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmdr9" event={"ID":"9df2fce8-7ad6-4572-8970-bcf915abc641","Type":"ContainerDied","Data":"3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427"} Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.113945 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmdr9" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.113974 4799 scope.go:117] "RemoveContainer" containerID="3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.113951 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmdr9" event={"ID":"9df2fce8-7ad6-4572-8970-bcf915abc641","Type":"ContainerDied","Data":"d63d89dd8d38a883a81f2f423c8a532f563fd4ab6e68eb58db913ac3b570bb57"} Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.147037 4799 scope.go:117] "RemoveContainer" containerID="d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.175037 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmdr9"] Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.177566 4799 scope.go:117] "RemoveContainer" containerID="48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.205335 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmdr9"] Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.248796 4799 scope.go:117] "RemoveContainer" containerID="3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427" Sep 30 15:17:07 crc kubenswrapper[4799]: E0930 15:17:07.249547 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427\": container with ID starting with 3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427 not found: ID does not exist" containerID="3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.249637 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427"} err="failed to get container status \"3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427\": rpc error: code = NotFound desc = could not find container \"3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427\": container with ID starting with 3c0c31768362ceade9c275f18dcda23c0d43eaebc318fb5c64d8b8d04c3c0427 not found: ID does not exist" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.249724 4799 scope.go:117] "RemoveContainer" containerID="d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e" Sep 30 15:17:07 crc kubenswrapper[4799]: E0930 15:17:07.250401 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e\": container with ID starting with d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e not found: ID does not exist" containerID="d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.250440 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e"} err="failed to get container status \"d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e\": rpc error: code = NotFound desc = could not find container \"d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e\": container with ID starting with d02f5e08b4555301c04a8e538102743ebeeb0d1f27d6f123f9c0d0ab8da7841e not found: ID does not exist" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.250481 4799 scope.go:117] "RemoveContainer" containerID="48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2" Sep 30 15:17:07 crc kubenswrapper[4799]: E0930 15:17:07.250783 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2\": container with ID starting with 48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2 not found: ID does not exist" containerID="48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2" Sep 30 15:17:07 crc kubenswrapper[4799]: I0930 15:17:07.250821 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2"} err="failed to get container status \"48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2\": rpc error: code = NotFound desc = could not find container \"48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2\": container with ID starting with 48fdf2e9ddfa5df0dd9f7c5b4bf4c62f91c36bee43a115f29ef85b19fb5e2ed2 not found: ID does not exist" Sep 30 15:17:08 crc kubenswrapper[4799]: I0930 15:17:08.524729 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" path="/var/lib/kubelet/pods/9df2fce8-7ad6-4572-8970-bcf915abc641/volumes" Sep 30 15:17:29 crc kubenswrapper[4799]: I0930 15:17:29.648452 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:17:29 crc kubenswrapper[4799]: I0930 15:17:29.649098 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:17:29 crc kubenswrapper[4799]: I0930 15:17:29.649171 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:17:29 crc kubenswrapper[4799]: I0930 15:17:29.650255 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"12492a0f5c49ea303b04b6e5eff4f9f2b1a2ef113c159201aca5e1132a65a941"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:17:29 crc kubenswrapper[4799]: I0930 15:17:29.650331 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://12492a0f5c49ea303b04b6e5eff4f9f2b1a2ef113c159201aca5e1132a65a941" gracePeriod=600 Sep 30 15:17:30 crc kubenswrapper[4799]: I0930 15:17:30.390883 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="12492a0f5c49ea303b04b6e5eff4f9f2b1a2ef113c159201aca5e1132a65a941" exitCode=0 Sep 30 15:17:30 crc kubenswrapper[4799]: I0930 15:17:30.390986 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"12492a0f5c49ea303b04b6e5eff4f9f2b1a2ef113c159201aca5e1132a65a941"} Sep 30 15:17:30 crc kubenswrapper[4799]: I0930 15:17:30.391309 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e"} Sep 30 15:17:30 crc kubenswrapper[4799]: I0930 15:17:30.391345 4799 scope.go:117] "RemoveContainer" containerID="cb345ef0f6396136be3167f08a30c5d535fdc41c62e54f483791c6c7529edf53" Sep 30 15:19:29 crc kubenswrapper[4799]: I0930 15:19:29.648799 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:19:29 crc kubenswrapper[4799]: I0930 15:19:29.651047 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:19:59 crc kubenswrapper[4799]: I0930 15:19:59.648379 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:19:59 crc kubenswrapper[4799]: I0930 15:19:59.649154 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:20:29 crc kubenswrapper[4799]: I0930 15:20:29.648809 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:20:29 crc kubenswrapper[4799]: I0930 15:20:29.649557 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:20:29 crc kubenswrapper[4799]: I0930 15:20:29.649629 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:20:29 crc kubenswrapper[4799]: I0930 15:20:29.650958 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:20:29 crc kubenswrapper[4799]: I0930 15:20:29.651099 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" gracePeriod=600 Sep 30 15:20:29 crc kubenswrapper[4799]: E0930 15:20:29.812093 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:20:30 crc kubenswrapper[4799]: I0930 15:20:30.451232 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" exitCode=0 Sep 30 15:20:30 crc kubenswrapper[4799]: I0930 15:20:30.451303 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e"} Sep 30 15:20:30 crc kubenswrapper[4799]: I0930 15:20:30.451381 4799 scope.go:117] "RemoveContainer" containerID="12492a0f5c49ea303b04b6e5eff4f9f2b1a2ef113c159201aca5e1132a65a941" Sep 30 15:20:30 crc kubenswrapper[4799]: I0930 15:20:30.452451 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:20:30 crc kubenswrapper[4799]: E0930 15:20:30.452880 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:20:42 crc kubenswrapper[4799]: I0930 15:20:42.504515 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:20:42 crc kubenswrapper[4799]: E0930 15:20:42.506184 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:20:57 crc kubenswrapper[4799]: I0930 15:20:57.503679 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:20:57 crc kubenswrapper[4799]: E0930 15:20:57.504759 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:21:10 crc kubenswrapper[4799]: I0930 15:21:10.503911 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:21:10 crc kubenswrapper[4799]: E0930 15:21:10.506138 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:21:24 crc kubenswrapper[4799]: I0930 15:21:24.504912 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:21:24 crc kubenswrapper[4799]: E0930 15:21:24.505924 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:21:36 crc kubenswrapper[4799]: I0930 15:21:36.503395 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:21:36 crc kubenswrapper[4799]: E0930 15:21:36.504385 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:21:50 crc kubenswrapper[4799]: I0930 15:21:50.503915 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:21:50 crc kubenswrapper[4799]: E0930 15:21:50.504935 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:22:04 crc kubenswrapper[4799]: I0930 15:22:04.504667 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:22:04 crc kubenswrapper[4799]: E0930 15:22:04.505531 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:22:15 crc kubenswrapper[4799]: I0930 15:22:15.503409 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:22:15 crc kubenswrapper[4799]: E0930 15:22:15.504235 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:22:30 crc kubenswrapper[4799]: I0930 15:22:30.504390 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:22:30 crc kubenswrapper[4799]: E0930 15:22:30.505369 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:22:43 crc kubenswrapper[4799]: I0930 15:22:43.503920 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:22:43 crc kubenswrapper[4799]: E0930 15:22:43.504802 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:22:54 crc kubenswrapper[4799]: I0930 15:22:54.504580 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:22:54 crc kubenswrapper[4799]: E0930 15:22:54.505488 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.638075 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sl6qg"] Sep 30 15:22:57 crc kubenswrapper[4799]: E0930 15:22:57.639197 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerName="registry-server" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.639215 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerName="registry-server" Sep 30 15:22:57 crc kubenswrapper[4799]: E0930 15:22:57.639272 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerName="extract-utilities" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.639280 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerName="extract-utilities" Sep 30 15:22:57 crc kubenswrapper[4799]: E0930 15:22:57.639296 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerName="extract-content" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.639304 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerName="extract-content" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.639547 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df2fce8-7ad6-4572-8970-bcf915abc641" containerName="registry-server" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.641569 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.646733 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sl6qg"] Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.745146 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-utilities\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.745585 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsgc6\" (UniqueName: \"kubernetes.io/projected/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-kube-api-access-dsgc6\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.745747 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-catalog-content\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.847976 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-catalog-content\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.848153 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-utilities\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.848187 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsgc6\" (UniqueName: \"kubernetes.io/projected/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-kube-api-access-dsgc6\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.848770 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-catalog-content\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.848784 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-utilities\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:57 crc kubenswrapper[4799]: I0930 15:22:57.875507 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsgc6\" (UniqueName: \"kubernetes.io/projected/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-kube-api-access-dsgc6\") pod \"certified-operators-sl6qg\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:58 crc kubenswrapper[4799]: I0930 15:22:58.018014 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:22:58 crc kubenswrapper[4799]: I0930 15:22:58.655074 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sl6qg"] Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.134681 4799 generic.go:334] "Generic (PLEG): container finished" podID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerID="523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1" exitCode=0 Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.134928 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sl6qg" event={"ID":"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d","Type":"ContainerDied","Data":"523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1"} Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.135037 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sl6qg" event={"ID":"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d","Type":"ContainerStarted","Data":"14070693f25e53aaec69b159eadd7bb3fc617be676919493dd1b47225ede29a4"} Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.137340 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.427720 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p5fbv"] Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.436474 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.443779 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5fbv"] Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.499497 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-utilities\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.499608 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-catalog-content\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.499841 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szwk4\" (UniqueName: \"kubernetes.io/projected/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-kube-api-access-szwk4\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.601338 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szwk4\" (UniqueName: \"kubernetes.io/projected/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-kube-api-access-szwk4\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.601959 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-utilities\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.602012 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-catalog-content\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.602620 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-catalog-content\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.602620 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-utilities\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.625724 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szwk4\" (UniqueName: \"kubernetes.io/projected/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-kube-api-access-szwk4\") pod \"community-operators-p5fbv\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:22:59 crc kubenswrapper[4799]: I0930 15:22:59.775681 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:23:00 crc kubenswrapper[4799]: I0930 15:23:00.342911 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5fbv"] Sep 30 15:23:01 crc kubenswrapper[4799]: I0930 15:23:01.168801 4799 generic.go:334] "Generic (PLEG): container finished" podID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerID="87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15" exitCode=0 Sep 30 15:23:01 crc kubenswrapper[4799]: I0930 15:23:01.169226 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fbv" event={"ID":"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5","Type":"ContainerDied","Data":"87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15"} Sep 30 15:23:01 crc kubenswrapper[4799]: I0930 15:23:01.169290 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fbv" event={"ID":"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5","Type":"ContainerStarted","Data":"e2d567cbdc2452fe45a269c279d84dc1a840e56ae4140751baf83144ce69e57b"} Sep 30 15:23:02 crc kubenswrapper[4799]: I0930 15:23:02.189978 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sl6qg" event={"ID":"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d","Type":"ContainerStarted","Data":"7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da"} Sep 30 15:23:03 crc kubenswrapper[4799]: I0930 15:23:03.202760 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fbv" event={"ID":"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5","Type":"ContainerStarted","Data":"8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100"} Sep 30 15:23:05 crc kubenswrapper[4799]: I0930 15:23:05.244378 4799 generic.go:334] "Generic (PLEG): container finished" podID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerID="7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da" exitCode=0 Sep 30 15:23:05 crc kubenswrapper[4799]: I0930 15:23:05.245195 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sl6qg" event={"ID":"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d","Type":"ContainerDied","Data":"7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da"} Sep 30 15:23:06 crc kubenswrapper[4799]: I0930 15:23:06.266829 4799 generic.go:334] "Generic (PLEG): container finished" podID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerID="8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100" exitCode=0 Sep 30 15:23:06 crc kubenswrapper[4799]: I0930 15:23:06.266933 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fbv" event={"ID":"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5","Type":"ContainerDied","Data":"8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100"} Sep 30 15:23:07 crc kubenswrapper[4799]: I0930 15:23:07.309150 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sl6qg" event={"ID":"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d","Type":"ContainerStarted","Data":"68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521"} Sep 30 15:23:07 crc kubenswrapper[4799]: I0930 15:23:07.318997 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fbv" event={"ID":"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5","Type":"ContainerStarted","Data":"90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1"} Sep 30 15:23:07 crc kubenswrapper[4799]: I0930 15:23:07.334552 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sl6qg" podStartSLOduration=3.423489395 podStartE2EDuration="10.334522432s" podCreationTimestamp="2025-09-30 15:22:57 +0000 UTC" firstStartedPulling="2025-09-30 15:22:59.137056054 +0000 UTC m=+3801.220656481" lastFinishedPulling="2025-09-30 15:23:06.048089091 +0000 UTC m=+3808.131689518" observedRunningTime="2025-09-30 15:23:07.331255216 +0000 UTC m=+3809.414855653" watchObservedRunningTime="2025-09-30 15:23:07.334522432 +0000 UTC m=+3809.418122859" Sep 30 15:23:07 crc kubenswrapper[4799]: I0930 15:23:07.363358 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p5fbv" podStartSLOduration=2.826494845 podStartE2EDuration="8.363326014s" podCreationTimestamp="2025-09-30 15:22:59 +0000 UTC" firstStartedPulling="2025-09-30 15:23:01.171414382 +0000 UTC m=+3803.255014799" lastFinishedPulling="2025-09-30 15:23:06.708245541 +0000 UTC m=+3808.791845968" observedRunningTime="2025-09-30 15:23:07.355094603 +0000 UTC m=+3809.438695030" watchObservedRunningTime="2025-09-30 15:23:07.363326014 +0000 UTC m=+3809.446926451" Sep 30 15:23:08 crc kubenswrapper[4799]: I0930 15:23:08.019267 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:23:08 crc kubenswrapper[4799]: I0930 15:23:08.019348 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:23:08 crc kubenswrapper[4799]: I0930 15:23:08.511770 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:23:08 crc kubenswrapper[4799]: E0930 15:23:08.526436 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:23:09 crc kubenswrapper[4799]: I0930 15:23:09.082862 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-sl6qg" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="registry-server" probeResult="failure" output=< Sep 30 15:23:09 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:23:09 crc kubenswrapper[4799]: > Sep 30 15:23:09 crc kubenswrapper[4799]: I0930 15:23:09.776827 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:23:09 crc kubenswrapper[4799]: I0930 15:23:09.777246 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:23:10 crc kubenswrapper[4799]: I0930 15:23:10.846850 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-p5fbv" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="registry-server" probeResult="failure" output=< Sep 30 15:23:10 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:23:10 crc kubenswrapper[4799]: > Sep 30 15:23:19 crc kubenswrapper[4799]: I0930 15:23:19.069520 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-sl6qg" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="registry-server" probeResult="failure" output=< Sep 30 15:23:19 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:23:19 crc kubenswrapper[4799]: > Sep 30 15:23:19 crc kubenswrapper[4799]: I0930 15:23:19.830717 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:23:19 crc kubenswrapper[4799]: I0930 15:23:19.891379 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:23:20 crc kubenswrapper[4799]: I0930 15:23:20.073585 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5fbv"] Sep 30 15:23:21 crc kubenswrapper[4799]: I0930 15:23:21.475623 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p5fbv" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="registry-server" containerID="cri-o://90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1" gracePeriod=2 Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.179483 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.283175 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-utilities\") pod \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.283433 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szwk4\" (UniqueName: \"kubernetes.io/projected/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-kube-api-access-szwk4\") pod \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.283461 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-catalog-content\") pod \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\" (UID: \"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5\") " Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.284475 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-utilities" (OuterVolumeSpecName: "utilities") pod "f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" (UID: "f112f2da-4e64-45c5-abfb-2d0f2d7e40d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.284900 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.301744 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-kube-api-access-szwk4" (OuterVolumeSpecName: "kube-api-access-szwk4") pod "f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" (UID: "f112f2da-4e64-45c5-abfb-2d0f2d7e40d5"). InnerVolumeSpecName "kube-api-access-szwk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.352569 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" (UID: "f112f2da-4e64-45c5-abfb-2d0f2d7e40d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.387108 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szwk4\" (UniqueName: \"kubernetes.io/projected/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-kube-api-access-szwk4\") on node \"crc\" DevicePath \"\"" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.387180 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.487853 4799 generic.go:334] "Generic (PLEG): container finished" podID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerID="90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1" exitCode=0 Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.487918 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fbv" event={"ID":"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5","Type":"ContainerDied","Data":"90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1"} Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.487964 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fbv" event={"ID":"f112f2da-4e64-45c5-abfb-2d0f2d7e40d5","Type":"ContainerDied","Data":"e2d567cbdc2452fe45a269c279d84dc1a840e56ae4140751baf83144ce69e57b"} Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.487985 4799 scope.go:117] "RemoveContainer" containerID="90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.487986 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5fbv" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.563854 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5fbv"] Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.564550 4799 scope.go:117] "RemoveContainer" containerID="8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.566904 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p5fbv"] Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.633662 4799 scope.go:117] "RemoveContainer" containerID="87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.677185 4799 scope.go:117] "RemoveContainer" containerID="90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1" Sep 30 15:23:22 crc kubenswrapper[4799]: E0930 15:23:22.678207 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1\": container with ID starting with 90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1 not found: ID does not exist" containerID="90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.678249 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1"} err="failed to get container status \"90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1\": rpc error: code = NotFound desc = could not find container \"90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1\": container with ID starting with 90b799258f7cf60a240591e3c7458ddacf4899ef02146f71c18b2a9f2c1003f1 not found: ID does not exist" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.678277 4799 scope.go:117] "RemoveContainer" containerID="8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100" Sep 30 15:23:22 crc kubenswrapper[4799]: E0930 15:23:22.678809 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100\": container with ID starting with 8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100 not found: ID does not exist" containerID="8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.678867 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100"} err="failed to get container status \"8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100\": rpc error: code = NotFound desc = could not find container \"8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100\": container with ID starting with 8b55a7a198dfad7a04785bb1263204c86e4d8122d2c410ca8d56b0210b133100 not found: ID does not exist" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.678901 4799 scope.go:117] "RemoveContainer" containerID="87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15" Sep 30 15:23:22 crc kubenswrapper[4799]: E0930 15:23:22.679221 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15\": container with ID starting with 87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15 not found: ID does not exist" containerID="87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15" Sep 30 15:23:22 crc kubenswrapper[4799]: I0930 15:23:22.679260 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15"} err="failed to get container status \"87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15\": rpc error: code = NotFound desc = could not find container \"87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15\": container with ID starting with 87b85e990e704dac13db6ba8a767f01fce817e2786877d2ee9f6c72bdf27da15 not found: ID does not exist" Sep 30 15:23:23 crc kubenswrapper[4799]: I0930 15:23:23.504028 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:23:23 crc kubenswrapper[4799]: E0930 15:23:23.504443 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:23:24 crc kubenswrapper[4799]: I0930 15:23:24.515009 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" path="/var/lib/kubelet/pods/f112f2da-4e64-45c5-abfb-2d0f2d7e40d5/volumes" Sep 30 15:23:28 crc kubenswrapper[4799]: I0930 15:23:28.081498 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:23:28 crc kubenswrapper[4799]: I0930 15:23:28.152385 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:23:28 crc kubenswrapper[4799]: I0930 15:23:28.835613 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sl6qg"] Sep 30 15:23:29 crc kubenswrapper[4799]: I0930 15:23:29.557554 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sl6qg" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="registry-server" containerID="cri-o://68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521" gracePeriod=2 Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.119533 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.270162 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsgc6\" (UniqueName: \"kubernetes.io/projected/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-kube-api-access-dsgc6\") pod \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.270255 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-catalog-content\") pod \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.270274 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-utilities\") pod \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\" (UID: \"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d\") " Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.271855 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-utilities" (OuterVolumeSpecName: "utilities") pod "7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" (UID: "7c3134dd-7ed5-4966-a02a-3963b8cf4a3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.277345 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-kube-api-access-dsgc6" (OuterVolumeSpecName: "kube-api-access-dsgc6") pod "7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" (UID: "7c3134dd-7ed5-4966-a02a-3963b8cf4a3d"). InnerVolumeSpecName "kube-api-access-dsgc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.341196 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" (UID: "7c3134dd-7ed5-4966-a02a-3963b8cf4a3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.372986 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsgc6\" (UniqueName: \"kubernetes.io/projected/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-kube-api-access-dsgc6\") on node \"crc\" DevicePath \"\"" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.373033 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.373049 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.572174 4799 generic.go:334] "Generic (PLEG): container finished" podID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerID="68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521" exitCode=0 Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.572234 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sl6qg" event={"ID":"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d","Type":"ContainerDied","Data":"68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521"} Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.572281 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sl6qg" event={"ID":"7c3134dd-7ed5-4966-a02a-3963b8cf4a3d","Type":"ContainerDied","Data":"14070693f25e53aaec69b159eadd7bb3fc617be676919493dd1b47225ede29a4"} Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.572315 4799 scope.go:117] "RemoveContainer" containerID="68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.573753 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sl6qg" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.608492 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sl6qg"] Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.608747 4799 scope.go:117] "RemoveContainer" containerID="7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.619678 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sl6qg"] Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.636487 4799 scope.go:117] "RemoveContainer" containerID="523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.724546 4799 scope.go:117] "RemoveContainer" containerID="68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521" Sep 30 15:23:30 crc kubenswrapper[4799]: E0930 15:23:30.725440 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521\": container with ID starting with 68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521 not found: ID does not exist" containerID="68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.725487 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521"} err="failed to get container status \"68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521\": rpc error: code = NotFound desc = could not find container \"68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521\": container with ID starting with 68c2bcd3f72fea62d64a37db06a22a184dccb2f894cfea148b48ab3bb9567521 not found: ID does not exist" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.725520 4799 scope.go:117] "RemoveContainer" containerID="7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da" Sep 30 15:23:30 crc kubenswrapper[4799]: E0930 15:23:30.725802 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da\": container with ID starting with 7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da not found: ID does not exist" containerID="7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.725871 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da"} err="failed to get container status \"7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da\": rpc error: code = NotFound desc = could not find container \"7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da\": container with ID starting with 7befe97b7780a97b3635b1b9439da8fa3a1660b71ad870d7087a62304bd769da not found: ID does not exist" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.725896 4799 scope.go:117] "RemoveContainer" containerID="523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1" Sep 30 15:23:30 crc kubenswrapper[4799]: E0930 15:23:30.726738 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1\": container with ID starting with 523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1 not found: ID does not exist" containerID="523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1" Sep 30 15:23:30 crc kubenswrapper[4799]: I0930 15:23:30.726766 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1"} err="failed to get container status \"523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1\": rpc error: code = NotFound desc = could not find container \"523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1\": container with ID starting with 523af94737d3253d431657cf03b650514c0a8a3bfae8ff3ec216dffaf4700ba1 not found: ID does not exist" Sep 30 15:23:32 crc kubenswrapper[4799]: I0930 15:23:32.519818 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" path="/var/lib/kubelet/pods/7c3134dd-7ed5-4966-a02a-3963b8cf4a3d/volumes" Sep 30 15:23:35 crc kubenswrapper[4799]: I0930 15:23:35.503602 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:23:35 crc kubenswrapper[4799]: E0930 15:23:35.504590 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:23:50 crc kubenswrapper[4799]: I0930 15:23:50.503726 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:23:50 crc kubenswrapper[4799]: E0930 15:23:50.504677 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:24:01 crc kubenswrapper[4799]: I0930 15:24:01.503504 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:24:01 crc kubenswrapper[4799]: E0930 15:24:01.504356 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:24:15 crc kubenswrapper[4799]: I0930 15:24:15.504499 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:24:15 crc kubenswrapper[4799]: E0930 15:24:15.506492 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:24:30 crc kubenswrapper[4799]: I0930 15:24:30.507075 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:24:30 crc kubenswrapper[4799]: E0930 15:24:30.507851 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:24:42 crc kubenswrapper[4799]: I0930 15:24:42.503523 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:24:42 crc kubenswrapper[4799]: E0930 15:24:42.504480 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:24:53 crc kubenswrapper[4799]: I0930 15:24:53.504328 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:24:53 crc kubenswrapper[4799]: E0930 15:24:53.505034 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:25:05 crc kubenswrapper[4799]: I0930 15:25:05.503266 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:25:05 crc kubenswrapper[4799]: E0930 15:25:05.504061 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:25:18 crc kubenswrapper[4799]: I0930 15:25:18.510968 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:25:18 crc kubenswrapper[4799]: E0930 15:25:18.511847 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:25:29 crc kubenswrapper[4799]: I0930 15:25:29.503459 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:25:29 crc kubenswrapper[4799]: E0930 15:25:29.504247 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:25:44 crc kubenswrapper[4799]: I0930 15:25:44.503289 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:25:44 crc kubenswrapper[4799]: I0930 15:25:44.913160 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"9b22b55685dd289c32676a886cdc7691beda3d8432f94ad8c274ec96fdf45de6"} Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.681637 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sfssn"] Sep 30 15:26:47 crc kubenswrapper[4799]: E0930 15:26:47.683598 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="extract-content" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.683736 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="extract-content" Sep 30 15:26:47 crc kubenswrapper[4799]: E0930 15:26:47.683809 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="extract-content" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.683898 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="extract-content" Sep 30 15:26:47 crc kubenswrapper[4799]: E0930 15:26:47.683980 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="extract-utilities" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.684038 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="extract-utilities" Sep 30 15:26:47 crc kubenswrapper[4799]: E0930 15:26:47.684099 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="registry-server" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.684161 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="registry-server" Sep 30 15:26:47 crc kubenswrapper[4799]: E0930 15:26:47.684225 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="extract-utilities" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.684281 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="extract-utilities" Sep 30 15:26:47 crc kubenswrapper[4799]: E0930 15:26:47.684344 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="registry-server" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.684406 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="registry-server" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.684695 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f112f2da-4e64-45c5-abfb-2d0f2d7e40d5" containerName="registry-server" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.684811 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3134dd-7ed5-4966-a02a-3963b8cf4a3d" containerName="registry-server" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.697584 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.792709 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de974362-13ef-46e4-9cc2-bc69f6b2db91-utilities\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.792977 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de974362-13ef-46e4-9cc2-bc69f6b2db91-catalog-content\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.793037 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg7bv\" (UniqueName: \"kubernetes.io/projected/de974362-13ef-46e4-9cc2-bc69f6b2db91-kube-api-access-lg7bv\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.804935 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sfssn"] Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.895668 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de974362-13ef-46e4-9cc2-bc69f6b2db91-catalog-content\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.895741 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg7bv\" (UniqueName: \"kubernetes.io/projected/de974362-13ef-46e4-9cc2-bc69f6b2db91-kube-api-access-lg7bv\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.895888 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de974362-13ef-46e4-9cc2-bc69f6b2db91-utilities\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.896471 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de974362-13ef-46e4-9cc2-bc69f6b2db91-utilities\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.896840 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de974362-13ef-46e4-9cc2-bc69f6b2db91-catalog-content\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:47 crc kubenswrapper[4799]: I0930 15:26:47.922919 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg7bv\" (UniqueName: \"kubernetes.io/projected/de974362-13ef-46e4-9cc2-bc69f6b2db91-kube-api-access-lg7bv\") pod \"redhat-operators-sfssn\" (UID: \"de974362-13ef-46e4-9cc2-bc69f6b2db91\") " pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:48 crc kubenswrapper[4799]: I0930 15:26:48.080320 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:26:48 crc kubenswrapper[4799]: I0930 15:26:48.682685 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sfssn"] Sep 30 15:26:49 crc kubenswrapper[4799]: I0930 15:26:49.550827 4799 generic.go:334] "Generic (PLEG): container finished" podID="de974362-13ef-46e4-9cc2-bc69f6b2db91" containerID="04e93b5cdb2a2ff2b8cf2b666ac2f31ff8a617ed5b5d0a7f7067920929aa6fc1" exitCode=0 Sep 30 15:26:49 crc kubenswrapper[4799]: I0930 15:26:49.551116 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfssn" event={"ID":"de974362-13ef-46e4-9cc2-bc69f6b2db91","Type":"ContainerDied","Data":"04e93b5cdb2a2ff2b8cf2b666ac2f31ff8a617ed5b5d0a7f7067920929aa6fc1"} Sep 30 15:26:49 crc kubenswrapper[4799]: I0930 15:26:49.551149 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfssn" event={"ID":"de974362-13ef-46e4-9cc2-bc69f6b2db91","Type":"ContainerStarted","Data":"63ab39c4b906425baf0b7940aae554847250e19ba53a30a771bfbe3f473794c5"} Sep 30 15:27:01 crc kubenswrapper[4799]: I0930 15:27:01.889666 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tmrfd"] Sep 30 15:27:01 crc kubenswrapper[4799]: I0930 15:27:01.892379 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:01 crc kubenswrapper[4799]: I0930 15:27:01.897629 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmrfd"] Sep 30 15:27:01 crc kubenswrapper[4799]: I0930 15:27:01.906121 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qjlt\" (UniqueName: \"kubernetes.io/projected/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-kube-api-access-9qjlt\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:01 crc kubenswrapper[4799]: I0930 15:27:01.906254 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-utilities\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:01 crc kubenswrapper[4799]: I0930 15:27:01.906276 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-catalog-content\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:02 crc kubenswrapper[4799]: I0930 15:27:02.007734 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-utilities\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:02 crc kubenswrapper[4799]: I0930 15:27:02.007791 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-catalog-content\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:02 crc kubenswrapper[4799]: I0930 15:27:02.007940 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qjlt\" (UniqueName: \"kubernetes.io/projected/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-kube-api-access-9qjlt\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:02 crc kubenswrapper[4799]: I0930 15:27:02.008310 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-catalog-content\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:02 crc kubenswrapper[4799]: I0930 15:27:02.008327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-utilities\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:02 crc kubenswrapper[4799]: I0930 15:27:02.030716 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qjlt\" (UniqueName: \"kubernetes.io/projected/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-kube-api-access-9qjlt\") pod \"redhat-marketplace-tmrfd\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:02 crc kubenswrapper[4799]: I0930 15:27:02.211448 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:04 crc kubenswrapper[4799]: E0930 15:27:04.578376 4799 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Sep 30 15:27:04 crc kubenswrapper[4799]: E0930 15:27:04.579137 4799 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lg7bv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-sfssn_openshift-marketplace(de974362-13ef-46e4-9cc2-bc69f6b2db91): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Sep 30 15:27:04 crc kubenswrapper[4799]: E0930 15:27:04.580551 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-sfssn" podUID="de974362-13ef-46e4-9cc2-bc69f6b2db91" Sep 30 15:27:04 crc kubenswrapper[4799]: I0930 15:27:04.714432 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmrfd"] Sep 30 15:27:04 crc kubenswrapper[4799]: E0930 15:27:04.715962 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-sfssn" podUID="de974362-13ef-46e4-9cc2-bc69f6b2db91" Sep 30 15:27:05 crc kubenswrapper[4799]: I0930 15:27:05.720851 4799 generic.go:334] "Generic (PLEG): container finished" podID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerID="f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6" exitCode=0 Sep 30 15:27:05 crc kubenswrapper[4799]: I0930 15:27:05.720955 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmrfd" event={"ID":"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d","Type":"ContainerDied","Data":"f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6"} Sep 30 15:27:05 crc kubenswrapper[4799]: I0930 15:27:05.721204 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmrfd" event={"ID":"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d","Type":"ContainerStarted","Data":"aa65f031df460e72c57697ffcf4812e23be7dd29b98c705f94708a6758487272"} Sep 30 15:27:07 crc kubenswrapper[4799]: I0930 15:27:07.740776 4799 generic.go:334] "Generic (PLEG): container finished" podID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerID="71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06" exitCode=0 Sep 30 15:27:07 crc kubenswrapper[4799]: I0930 15:27:07.740985 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmrfd" event={"ID":"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d","Type":"ContainerDied","Data":"71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06"} Sep 30 15:27:07 crc kubenswrapper[4799]: I0930 15:27:07.889337 4799 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5cfc9f87d7-nj7t8" podUID="f32c852d-173a-4417-a268-38b7a480937e" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Sep 30 15:27:09 crc kubenswrapper[4799]: I0930 15:27:09.780893 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmrfd" event={"ID":"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d","Type":"ContainerStarted","Data":"f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720"} Sep 30 15:27:09 crc kubenswrapper[4799]: I0930 15:27:09.812336 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tmrfd" podStartSLOduration=6.065223623 podStartE2EDuration="8.812313423s" podCreationTimestamp="2025-09-30 15:27:01 +0000 UTC" firstStartedPulling="2025-09-30 15:27:05.722747827 +0000 UTC m=+4047.806348254" lastFinishedPulling="2025-09-30 15:27:08.469837627 +0000 UTC m=+4050.553438054" observedRunningTime="2025-09-30 15:27:09.805094842 +0000 UTC m=+4051.888695269" watchObservedRunningTime="2025-09-30 15:27:09.812313423 +0000 UTC m=+4051.895913840" Sep 30 15:27:12 crc kubenswrapper[4799]: I0930 15:27:12.212011 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:12 crc kubenswrapper[4799]: I0930 15:27:12.212572 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:12 crc kubenswrapper[4799]: I0930 15:27:12.266597 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:19 crc kubenswrapper[4799]: I0930 15:27:19.878324 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfssn" event={"ID":"de974362-13ef-46e4-9cc2-bc69f6b2db91","Type":"ContainerStarted","Data":"49c0939eca0a6a5da9b5d91f20fa5835d93731feebee99c2b75648d3f676d6f5"} Sep 30 15:27:22 crc kubenswrapper[4799]: I0930 15:27:22.267640 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:22 crc kubenswrapper[4799]: I0930 15:27:22.365666 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmrfd"] Sep 30 15:27:22 crc kubenswrapper[4799]: I0930 15:27:22.909234 4799 generic.go:334] "Generic (PLEG): container finished" podID="de974362-13ef-46e4-9cc2-bc69f6b2db91" containerID="49c0939eca0a6a5da9b5d91f20fa5835d93731feebee99c2b75648d3f676d6f5" exitCode=0 Sep 30 15:27:22 crc kubenswrapper[4799]: I0930 15:27:22.909866 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tmrfd" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerName="registry-server" containerID="cri-o://f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720" gracePeriod=2 Sep 30 15:27:22 crc kubenswrapper[4799]: I0930 15:27:22.909313 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfssn" event={"ID":"de974362-13ef-46e4-9cc2-bc69f6b2db91","Type":"ContainerDied","Data":"49c0939eca0a6a5da9b5d91f20fa5835d93731feebee99c2b75648d3f676d6f5"} Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.464207 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.549412 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qjlt\" (UniqueName: \"kubernetes.io/projected/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-kube-api-access-9qjlt\") pod \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.549842 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-catalog-content\") pod \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.550049 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-utilities\") pod \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\" (UID: \"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d\") " Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.551402 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-utilities" (OuterVolumeSpecName: "utilities") pod "f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" (UID: "f8fc6ef5-4ed5-4823-bb09-581ed6bf512d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.556391 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-kube-api-access-9qjlt" (OuterVolumeSpecName: "kube-api-access-9qjlt") pod "f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" (UID: "f8fc6ef5-4ed5-4823-bb09-581ed6bf512d"). InnerVolumeSpecName "kube-api-access-9qjlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.571803 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" (UID: "f8fc6ef5-4ed5-4823-bb09-581ed6bf512d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.652745 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qjlt\" (UniqueName: \"kubernetes.io/projected/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-kube-api-access-9qjlt\") on node \"crc\" DevicePath \"\"" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.652799 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.652809 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.919947 4799 generic.go:334] "Generic (PLEG): container finished" podID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerID="f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720" exitCode=0 Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.920025 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmrfd" event={"ID":"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d","Type":"ContainerDied","Data":"f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720"} Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.920086 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tmrfd" event={"ID":"f8fc6ef5-4ed5-4823-bb09-581ed6bf512d","Type":"ContainerDied","Data":"aa65f031df460e72c57697ffcf4812e23be7dd29b98c705f94708a6758487272"} Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.920110 4799 scope.go:117] "RemoveContainer" containerID="f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.920471 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tmrfd" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.943568 4799 scope.go:117] "RemoveContainer" containerID="71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06" Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.961138 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmrfd"] Sep 30 15:27:23 crc kubenswrapper[4799]: I0930 15:27:23.968713 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tmrfd"] Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.093806 4799 scope.go:117] "RemoveContainer" containerID="f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6" Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.149965 4799 scope.go:117] "RemoveContainer" containerID="f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720" Sep 30 15:27:24 crc kubenswrapper[4799]: E0930 15:27:24.150833 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720\": container with ID starting with f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720 not found: ID does not exist" containerID="f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720" Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.150891 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720"} err="failed to get container status \"f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720\": rpc error: code = NotFound desc = could not find container \"f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720\": container with ID starting with f7380b90f3b14c32d13ea65725d856ae5f22f2d17fa122d0bbc07b633599a720 not found: ID does not exist" Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.150922 4799 scope.go:117] "RemoveContainer" containerID="71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06" Sep 30 15:27:24 crc kubenswrapper[4799]: E0930 15:27:24.151363 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06\": container with ID starting with 71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06 not found: ID does not exist" containerID="71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06" Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.151397 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06"} err="failed to get container status \"71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06\": rpc error: code = NotFound desc = could not find container \"71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06\": container with ID starting with 71e74886eca9ced8ab9cebe1aacff974a72f91037a5e5509faab10cc012a2d06 not found: ID does not exist" Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.151420 4799 scope.go:117] "RemoveContainer" containerID="f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6" Sep 30 15:27:24 crc kubenswrapper[4799]: E0930 15:27:24.151776 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6\": container with ID starting with f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6 not found: ID does not exist" containerID="f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6" Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.151811 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6"} err="failed to get container status \"f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6\": rpc error: code = NotFound desc = could not find container \"f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6\": container with ID starting with f5d8914ca71bff1ec43fad7059e512b8f8bd4f82b18ef7bb6c1736d35ec722f6 not found: ID does not exist" Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.515489 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" path="/var/lib/kubelet/pods/f8fc6ef5-4ed5-4823-bb09-581ed6bf512d/volumes" Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.932730 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sfssn" event={"ID":"de974362-13ef-46e4-9cc2-bc69f6b2db91","Type":"ContainerStarted","Data":"b464406e624c629bfba05f19e5cb4b4ace7b553e77bfac4b3a573b77dcc24b7b"} Sep 30 15:27:24 crc kubenswrapper[4799]: I0930 15:27:24.957407 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sfssn" podStartSLOduration=4.00151353 podStartE2EDuration="37.957379927s" podCreationTimestamp="2025-09-30 15:26:47 +0000 UTC" firstStartedPulling="2025-09-30 15:26:49.555555839 +0000 UTC m=+4031.639156266" lastFinishedPulling="2025-09-30 15:27:23.511422236 +0000 UTC m=+4065.595022663" observedRunningTime="2025-09-30 15:27:24.952982608 +0000 UTC m=+4067.036583055" watchObservedRunningTime="2025-09-30 15:27:24.957379927 +0000 UTC m=+4067.040980354" Sep 30 15:27:28 crc kubenswrapper[4799]: I0930 15:27:28.081853 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:27:28 crc kubenswrapper[4799]: I0930 15:27:28.082182 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:27:29 crc kubenswrapper[4799]: I0930 15:27:29.141544 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sfssn" podUID="de974362-13ef-46e4-9cc2-bc69f6b2db91" containerName="registry-server" probeResult="failure" output=< Sep 30 15:27:29 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:27:29 crc kubenswrapper[4799]: > Sep 30 15:27:38 crc kubenswrapper[4799]: I0930 15:27:38.130684 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:27:38 crc kubenswrapper[4799]: I0930 15:27:38.180942 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sfssn" Sep 30 15:27:38 crc kubenswrapper[4799]: I0930 15:27:38.263217 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sfssn"] Sep 30 15:27:38 crc kubenswrapper[4799]: I0930 15:27:38.375188 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tplx2"] Sep 30 15:27:38 crc kubenswrapper[4799]: I0930 15:27:38.375865 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tplx2" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerName="registry-server" containerID="cri-o://90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d" gracePeriod=2 Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.027965 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.065998 4799 generic.go:334] "Generic (PLEG): container finished" podID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerID="90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d" exitCode=0 Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.066085 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tplx2" event={"ID":"eb9f79d7-c4db-4f1f-9b99-6c577dca041d","Type":"ContainerDied","Data":"90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d"} Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.066116 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tplx2" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.066163 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tplx2" event={"ID":"eb9f79d7-c4db-4f1f-9b99-6c577dca041d","Type":"ContainerDied","Data":"bf8bfcd40b9a7fd5b501abe6baf79fe640957e9e27ea58b410cd4f9b8be4627b"} Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.066194 4799 scope.go:117] "RemoveContainer" containerID="90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.098311 4799 scope.go:117] "RemoveContainer" containerID="14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.131984 4799 scope.go:117] "RemoveContainer" containerID="2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.190500 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8gm4\" (UniqueName: \"kubernetes.io/projected/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-kube-api-access-r8gm4\") pod \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.190585 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-catalog-content\") pod \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.190624 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-utilities\") pod \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\" (UID: \"eb9f79d7-c4db-4f1f-9b99-6c577dca041d\") " Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.191630 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-utilities" (OuterVolumeSpecName: "utilities") pod "eb9f79d7-c4db-4f1f-9b99-6c577dca041d" (UID: "eb9f79d7-c4db-4f1f-9b99-6c577dca041d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.205707 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-kube-api-access-r8gm4" (OuterVolumeSpecName: "kube-api-access-r8gm4") pod "eb9f79d7-c4db-4f1f-9b99-6c577dca041d" (UID: "eb9f79d7-c4db-4f1f-9b99-6c577dca041d"). InnerVolumeSpecName "kube-api-access-r8gm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.216957 4799 scope.go:117] "RemoveContainer" containerID="90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d" Sep 30 15:27:39 crc kubenswrapper[4799]: E0930 15:27:39.219149 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d\": container with ID starting with 90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d not found: ID does not exist" containerID="90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.219191 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d"} err="failed to get container status \"90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d\": rpc error: code = NotFound desc = could not find container \"90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d\": container with ID starting with 90e212ea3b04dd0c4928105990d66b71756678050baaa37319b81c64728a577d not found: ID does not exist" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.219221 4799 scope.go:117] "RemoveContainer" containerID="14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb" Sep 30 15:27:39 crc kubenswrapper[4799]: E0930 15:27:39.219902 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb\": container with ID starting with 14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb not found: ID does not exist" containerID="14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.219924 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb"} err="failed to get container status \"14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb\": rpc error: code = NotFound desc = could not find container \"14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb\": container with ID starting with 14f3b8bb42a403ac432545379c4d78b4b69d5fb5c74707530ded8f96da5d3dfb not found: ID does not exist" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.219939 4799 scope.go:117] "RemoveContainer" containerID="2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f" Sep 30 15:27:39 crc kubenswrapper[4799]: E0930 15:27:39.224839 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f\": container with ID starting with 2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f not found: ID does not exist" containerID="2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.225085 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f"} err="failed to get container status \"2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f\": rpc error: code = NotFound desc = could not find container \"2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f\": container with ID starting with 2cee58ea5ceb0a50bda9d7abe14710d222c685043f80ab48fe6b10c2d3f38c8f not found: ID does not exist" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.271476 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb9f79d7-c4db-4f1f-9b99-6c577dca041d" (UID: "eb9f79d7-c4db-4f1f-9b99-6c577dca041d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.293636 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8gm4\" (UniqueName: \"kubernetes.io/projected/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-kube-api-access-r8gm4\") on node \"crc\" DevicePath \"\"" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.293905 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.293917 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb9f79d7-c4db-4f1f-9b99-6c577dca041d-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.457122 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tplx2"] Sep 30 15:27:39 crc kubenswrapper[4799]: I0930 15:27:39.472066 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tplx2"] Sep 30 15:27:40 crc kubenswrapper[4799]: I0930 15:27:40.516363 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" path="/var/lib/kubelet/pods/eb9f79d7-c4db-4f1f-9b99-6c577dca041d/volumes" Sep 30 15:27:59 crc kubenswrapper[4799]: I0930 15:27:59.648665 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:27:59 crc kubenswrapper[4799]: I0930 15:27:59.649237 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:28:29 crc kubenswrapper[4799]: I0930 15:28:29.648408 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:28:29 crc kubenswrapper[4799]: I0930 15:28:29.649876 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:28:59 crc kubenswrapper[4799]: I0930 15:28:59.648448 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:28:59 crc kubenswrapper[4799]: I0930 15:28:59.649041 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:28:59 crc kubenswrapper[4799]: I0930 15:28:59.649095 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:28:59 crc kubenswrapper[4799]: I0930 15:28:59.649906 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b22b55685dd289c32676a886cdc7691beda3d8432f94ad8c274ec96fdf45de6"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:28:59 crc kubenswrapper[4799]: I0930 15:28:59.649964 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://9b22b55685dd289c32676a886cdc7691beda3d8432f94ad8c274ec96fdf45de6" gracePeriod=600 Sep 30 15:28:59 crc kubenswrapper[4799]: I0930 15:28:59.822634 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="9b22b55685dd289c32676a886cdc7691beda3d8432f94ad8c274ec96fdf45de6" exitCode=0 Sep 30 15:28:59 crc kubenswrapper[4799]: I0930 15:28:59.822700 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"9b22b55685dd289c32676a886cdc7691beda3d8432f94ad8c274ec96fdf45de6"} Sep 30 15:28:59 crc kubenswrapper[4799]: I0930 15:28:59.823353 4799 scope.go:117] "RemoveContainer" containerID="597b7d679da6ffee99bd7c566791af3a53d1b4e16fda43c2723e9378d9134e4e" Sep 30 15:29:00 crc kubenswrapper[4799]: I0930 15:29:00.837500 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12"} Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.155305 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns"] Sep 30 15:30:00 crc kubenswrapper[4799]: E0930 15:30:00.156390 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerName="extract-utilities" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.156414 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerName="extract-utilities" Sep 30 15:30:00 crc kubenswrapper[4799]: E0930 15:30:00.156457 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerName="extract-content" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.156468 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerName="extract-content" Sep 30 15:30:00 crc kubenswrapper[4799]: E0930 15:30:00.156491 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerName="extract-content" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.156499 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerName="extract-content" Sep 30 15:30:00 crc kubenswrapper[4799]: E0930 15:30:00.156514 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerName="registry-server" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.156521 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerName="registry-server" Sep 30 15:30:00 crc kubenswrapper[4799]: E0930 15:30:00.156530 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerName="registry-server" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.156537 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerName="registry-server" Sep 30 15:30:00 crc kubenswrapper[4799]: E0930 15:30:00.156544 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerName="extract-utilities" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.156554 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerName="extract-utilities" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.156786 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9f79d7-c4db-4f1f-9b99-6c577dca041d" containerName="registry-server" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.156832 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8fc6ef5-4ed5-4823-bb09-581ed6bf512d" containerName="registry-server" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.159424 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.162249 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.168208 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns"] Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.169257 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.219876 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9hbw\" (UniqueName: \"kubernetes.io/projected/c4caac6a-24f5-4892-960c-2253b4c8bbb4-kube-api-access-m9hbw\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.219990 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4caac6a-24f5-4892-960c-2253b4c8bbb4-secret-volume\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.220086 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4caac6a-24f5-4892-960c-2253b4c8bbb4-config-volume\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.322328 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4caac6a-24f5-4892-960c-2253b4c8bbb4-config-volume\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.322489 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9hbw\" (UniqueName: \"kubernetes.io/projected/c4caac6a-24f5-4892-960c-2253b4c8bbb4-kube-api-access-m9hbw\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.322626 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4caac6a-24f5-4892-960c-2253b4c8bbb4-secret-volume\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.323725 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4caac6a-24f5-4892-960c-2253b4c8bbb4-config-volume\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.331945 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4caac6a-24f5-4892-960c-2253b4c8bbb4-secret-volume\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.343966 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9hbw\" (UniqueName: \"kubernetes.io/projected/c4caac6a-24f5-4892-960c-2253b4c8bbb4-kube-api-access-m9hbw\") pod \"collect-profiles-29320770-qh5ns\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.487750 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:00 crc kubenswrapper[4799]: I0930 15:30:00.983406 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns"] Sep 30 15:30:01 crc kubenswrapper[4799]: I0930 15:30:01.376976 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" event={"ID":"c4caac6a-24f5-4892-960c-2253b4c8bbb4","Type":"ContainerStarted","Data":"3a0c092ca1f1f6d639ee1f45beb1e5cbf0ec643ae7afcb7d9682858d7d97b1a9"} Sep 30 15:30:01 crc kubenswrapper[4799]: I0930 15:30:01.377051 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" event={"ID":"c4caac6a-24f5-4892-960c-2253b4c8bbb4","Type":"ContainerStarted","Data":"263db7445e216e42cd260493740376ada3f92f9eb06ce42f203b0e00c3022c18"} Sep 30 15:30:02 crc kubenswrapper[4799]: I0930 15:30:02.386993 4799 generic.go:334] "Generic (PLEG): container finished" podID="c4caac6a-24f5-4892-960c-2253b4c8bbb4" containerID="3a0c092ca1f1f6d639ee1f45beb1e5cbf0ec643ae7afcb7d9682858d7d97b1a9" exitCode=0 Sep 30 15:30:02 crc kubenswrapper[4799]: I0930 15:30:02.387350 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" event={"ID":"c4caac6a-24f5-4892-960c-2253b4c8bbb4","Type":"ContainerDied","Data":"3a0c092ca1f1f6d639ee1f45beb1e5cbf0ec643ae7afcb7d9682858d7d97b1a9"} Sep 30 15:30:03 crc kubenswrapper[4799]: I0930 15:30:03.801192 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:03 crc kubenswrapper[4799]: I0930 15:30:03.898709 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9hbw\" (UniqueName: \"kubernetes.io/projected/c4caac6a-24f5-4892-960c-2253b4c8bbb4-kube-api-access-m9hbw\") pod \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " Sep 30 15:30:03 crc kubenswrapper[4799]: I0930 15:30:03.898961 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4caac6a-24f5-4892-960c-2253b4c8bbb4-secret-volume\") pod \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " Sep 30 15:30:03 crc kubenswrapper[4799]: I0930 15:30:03.899008 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4caac6a-24f5-4892-960c-2253b4c8bbb4-config-volume\") pod \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\" (UID: \"c4caac6a-24f5-4892-960c-2253b4c8bbb4\") " Sep 30 15:30:03 crc kubenswrapper[4799]: I0930 15:30:03.899566 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4caac6a-24f5-4892-960c-2253b4c8bbb4-config-volume" (OuterVolumeSpecName: "config-volume") pod "c4caac6a-24f5-4892-960c-2253b4c8bbb4" (UID: "c4caac6a-24f5-4892-960c-2253b4c8bbb4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 15:30:03 crc kubenswrapper[4799]: I0930 15:30:03.909918 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4caac6a-24f5-4892-960c-2253b4c8bbb4-kube-api-access-m9hbw" (OuterVolumeSpecName: "kube-api-access-m9hbw") pod "c4caac6a-24f5-4892-960c-2253b4c8bbb4" (UID: "c4caac6a-24f5-4892-960c-2253b4c8bbb4"). InnerVolumeSpecName "kube-api-access-m9hbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:30:03 crc kubenswrapper[4799]: I0930 15:30:03.910056 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4caac6a-24f5-4892-960c-2253b4c8bbb4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c4caac6a-24f5-4892-960c-2253b4c8bbb4" (UID: "c4caac6a-24f5-4892-960c-2253b4c8bbb4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:30:04 crc kubenswrapper[4799]: I0930 15:30:04.001690 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4caac6a-24f5-4892-960c-2253b4c8bbb4-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:30:04 crc kubenswrapper[4799]: I0930 15:30:04.001730 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4caac6a-24f5-4892-960c-2253b4c8bbb4-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:30:04 crc kubenswrapper[4799]: I0930 15:30:04.001743 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9hbw\" (UniqueName: \"kubernetes.io/projected/c4caac6a-24f5-4892-960c-2253b4c8bbb4-kube-api-access-m9hbw\") on node \"crc\" DevicePath \"\"" Sep 30 15:30:04 crc kubenswrapper[4799]: I0930 15:30:04.410466 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" event={"ID":"c4caac6a-24f5-4892-960c-2253b4c8bbb4","Type":"ContainerDied","Data":"263db7445e216e42cd260493740376ada3f92f9eb06ce42f203b0e00c3022c18"} Sep 30 15:30:04 crc kubenswrapper[4799]: I0930 15:30:04.410766 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="263db7445e216e42cd260493740376ada3f92f9eb06ce42f203b0e00c3022c18" Sep 30 15:30:04 crc kubenswrapper[4799]: I0930 15:30:04.410531 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320770-qh5ns" Sep 30 15:30:04 crc kubenswrapper[4799]: I0930 15:30:04.891230 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49"] Sep 30 15:30:04 crc kubenswrapper[4799]: I0930 15:30:04.903154 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320725-p9j49"] Sep 30 15:30:06 crc kubenswrapper[4799]: I0930 15:30:06.518196 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e013782-a27d-4cf6-bc7c-29a13bbd4137" path="/var/lib/kubelet/pods/1e013782-a27d-4cf6-bc7c-29a13bbd4137/volumes" Sep 30 15:30:13 crc kubenswrapper[4799]: I0930 15:30:13.194559 4799 scope.go:117] "RemoveContainer" containerID="cb13ec04eb93bc6be8b6a7ea9c48dddfcbd4ca4e489636300c5b03a962ff2961" Sep 30 15:31:29 crc kubenswrapper[4799]: I0930 15:31:29.648925 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:31:29 crc kubenswrapper[4799]: I0930 15:31:29.649594 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:31:59 crc kubenswrapper[4799]: I0930 15:31:59.648571 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:31:59 crc kubenswrapper[4799]: I0930 15:31:59.649103 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.648541 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.649184 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.649244 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.650169 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.650269 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" gracePeriod=600 Sep 30 15:32:29 crc kubenswrapper[4799]: E0930 15:32:29.779465 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.806582 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" exitCode=0 Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.806638 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12"} Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.806756 4799 scope.go:117] "RemoveContainer" containerID="9b22b55685dd289c32676a886cdc7691beda3d8432f94ad8c274ec96fdf45de6" Sep 30 15:32:29 crc kubenswrapper[4799]: I0930 15:32:29.807529 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:32:29 crc kubenswrapper[4799]: E0930 15:32:29.807947 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:32:42 crc kubenswrapper[4799]: I0930 15:32:42.503130 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:32:42 crc kubenswrapper[4799]: E0930 15:32:42.503830 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:32:54 crc kubenswrapper[4799]: I0930 15:32:54.503686 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:32:54 crc kubenswrapper[4799]: E0930 15:32:54.504773 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:33:08 crc kubenswrapper[4799]: I0930 15:33:08.512450 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:33:08 crc kubenswrapper[4799]: E0930 15:33:08.514087 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:33:20 crc kubenswrapper[4799]: I0930 15:33:20.503138 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:33:20 crc kubenswrapper[4799]: E0930 15:33:20.504839 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:33:23 crc kubenswrapper[4799]: I0930 15:33:23.969905 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m4n54"] Sep 30 15:33:23 crc kubenswrapper[4799]: E0930 15:33:23.970797 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4caac6a-24f5-4892-960c-2253b4c8bbb4" containerName="collect-profiles" Sep 30 15:33:23 crc kubenswrapper[4799]: I0930 15:33:23.970810 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4caac6a-24f5-4892-960c-2253b4c8bbb4" containerName="collect-profiles" Sep 30 15:33:23 crc kubenswrapper[4799]: I0930 15:33:23.971015 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4caac6a-24f5-4892-960c-2253b4c8bbb4" containerName="collect-profiles" Sep 30 15:33:23 crc kubenswrapper[4799]: I0930 15:33:23.972476 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:23 crc kubenswrapper[4799]: I0930 15:33:23.985698 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m4n54"] Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.171740 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-catalog-content\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.171889 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-utilities\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.172056 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgjrq\" (UniqueName: \"kubernetes.io/projected/717cbf7e-4d14-4174-9de7-eba44c406e6e-kube-api-access-wgjrq\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.273705 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-utilities\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.273812 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgjrq\" (UniqueName: \"kubernetes.io/projected/717cbf7e-4d14-4174-9de7-eba44c406e6e-kube-api-access-wgjrq\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.273934 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-catalog-content\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.275012 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-catalog-content\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.275609 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-utilities\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.305523 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgjrq\" (UniqueName: \"kubernetes.io/projected/717cbf7e-4d14-4174-9de7-eba44c406e6e-kube-api-access-wgjrq\") pod \"certified-operators-m4n54\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:24 crc kubenswrapper[4799]: I0930 15:33:24.598416 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:25 crc kubenswrapper[4799]: I0930 15:33:25.344093 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m4n54"] Sep 30 15:33:26 crc kubenswrapper[4799]: I0930 15:33:26.357821 4799 generic.go:334] "Generic (PLEG): container finished" podID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerID="9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f" exitCode=0 Sep 30 15:33:26 crc kubenswrapper[4799]: I0930 15:33:26.358100 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4n54" event={"ID":"717cbf7e-4d14-4174-9de7-eba44c406e6e","Type":"ContainerDied","Data":"9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f"} Sep 30 15:33:26 crc kubenswrapper[4799]: I0930 15:33:26.358223 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4n54" event={"ID":"717cbf7e-4d14-4174-9de7-eba44c406e6e","Type":"ContainerStarted","Data":"8740a27af3190ad9a97bb1781cd1d86bfab41c7e6293759d7a97fff7206889c5"} Sep 30 15:33:26 crc kubenswrapper[4799]: I0930 15:33:26.360003 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 15:33:27 crc kubenswrapper[4799]: I0930 15:33:27.369351 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4n54" event={"ID":"717cbf7e-4d14-4174-9de7-eba44c406e6e","Type":"ContainerStarted","Data":"4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97"} Sep 30 15:33:28 crc kubenswrapper[4799]: I0930 15:33:28.381161 4799 generic.go:334] "Generic (PLEG): container finished" podID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerID="4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97" exitCode=0 Sep 30 15:33:28 crc kubenswrapper[4799]: I0930 15:33:28.381266 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4n54" event={"ID":"717cbf7e-4d14-4174-9de7-eba44c406e6e","Type":"ContainerDied","Data":"4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97"} Sep 30 15:33:29 crc kubenswrapper[4799]: I0930 15:33:29.392448 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4n54" event={"ID":"717cbf7e-4d14-4174-9de7-eba44c406e6e","Type":"ContainerStarted","Data":"eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a"} Sep 30 15:33:29 crc kubenswrapper[4799]: I0930 15:33:29.417289 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m4n54" podStartSLOduration=3.653841742 podStartE2EDuration="6.417263709s" podCreationTimestamp="2025-09-30 15:33:23 +0000 UTC" firstStartedPulling="2025-09-30 15:33:26.35966823 +0000 UTC m=+4428.443268657" lastFinishedPulling="2025-09-30 15:33:29.123090197 +0000 UTC m=+4431.206690624" observedRunningTime="2025-09-30 15:33:29.407983878 +0000 UTC m=+4431.491584335" watchObservedRunningTime="2025-09-30 15:33:29.417263709 +0000 UTC m=+4431.500864136" Sep 30 15:33:34 crc kubenswrapper[4799]: I0930 15:33:34.504677 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:33:34 crc kubenswrapper[4799]: E0930 15:33:34.505489 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:33:34 crc kubenswrapper[4799]: I0930 15:33:34.599519 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:34 crc kubenswrapper[4799]: I0930 15:33:34.599568 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:34 crc kubenswrapper[4799]: I0930 15:33:34.644680 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:35 crc kubenswrapper[4799]: I0930 15:33:35.716055 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:35 crc kubenswrapper[4799]: I0930 15:33:35.764789 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m4n54"] Sep 30 15:33:37 crc kubenswrapper[4799]: I0930 15:33:37.457356 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m4n54" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerName="registry-server" containerID="cri-o://eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a" gracePeriod=2 Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.096179 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.171370 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-catalog-content\") pod \"717cbf7e-4d14-4174-9de7-eba44c406e6e\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.171588 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-utilities\") pod \"717cbf7e-4d14-4174-9de7-eba44c406e6e\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.171754 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgjrq\" (UniqueName: \"kubernetes.io/projected/717cbf7e-4d14-4174-9de7-eba44c406e6e-kube-api-access-wgjrq\") pod \"717cbf7e-4d14-4174-9de7-eba44c406e6e\" (UID: \"717cbf7e-4d14-4174-9de7-eba44c406e6e\") " Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.173726 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-utilities" (OuterVolumeSpecName: "utilities") pod "717cbf7e-4d14-4174-9de7-eba44c406e6e" (UID: "717cbf7e-4d14-4174-9de7-eba44c406e6e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.198927 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/717cbf7e-4d14-4174-9de7-eba44c406e6e-kube-api-access-wgjrq" (OuterVolumeSpecName: "kube-api-access-wgjrq") pod "717cbf7e-4d14-4174-9de7-eba44c406e6e" (UID: "717cbf7e-4d14-4174-9de7-eba44c406e6e"). InnerVolumeSpecName "kube-api-access-wgjrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.274243 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.274279 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgjrq\" (UniqueName: \"kubernetes.io/projected/717cbf7e-4d14-4174-9de7-eba44c406e6e-kube-api-access-wgjrq\") on node \"crc\" DevicePath \"\"" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.488035 4799 generic.go:334] "Generic (PLEG): container finished" podID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerID="eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a" exitCode=0 Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.488117 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4n54" event={"ID":"717cbf7e-4d14-4174-9de7-eba44c406e6e","Type":"ContainerDied","Data":"eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a"} Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.488159 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m4n54" event={"ID":"717cbf7e-4d14-4174-9de7-eba44c406e6e","Type":"ContainerDied","Data":"8740a27af3190ad9a97bb1781cd1d86bfab41c7e6293759d7a97fff7206889c5"} Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.488194 4799 scope.go:117] "RemoveContainer" containerID="eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.488579 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m4n54" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.528766 4799 scope.go:117] "RemoveContainer" containerID="4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.577655 4799 scope.go:117] "RemoveContainer" containerID="9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.684752 4799 scope.go:117] "RemoveContainer" containerID="eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a" Sep 30 15:33:38 crc kubenswrapper[4799]: E0930 15:33:38.686493 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a\": container with ID starting with eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a not found: ID does not exist" containerID="eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.686697 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a"} err="failed to get container status \"eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a\": rpc error: code = NotFound desc = could not find container \"eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a\": container with ID starting with eb843d7784801b0b0f96b880461590bd5b7122ea059600e258ab867983fdc89a not found: ID does not exist" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.686861 4799 scope.go:117] "RemoveContainer" containerID="4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97" Sep 30 15:33:38 crc kubenswrapper[4799]: E0930 15:33:38.687323 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97\": container with ID starting with 4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97 not found: ID does not exist" containerID="4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.687445 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97"} err="failed to get container status \"4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97\": rpc error: code = NotFound desc = could not find container \"4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97\": container with ID starting with 4a3924c082dd3be358848b666a87121cc0108bf696ec19e91e82096d2cfc9c97 not found: ID does not exist" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.687543 4799 scope.go:117] "RemoveContainer" containerID="9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f" Sep 30 15:33:38 crc kubenswrapper[4799]: E0930 15:33:38.688122 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f\": container with ID starting with 9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f not found: ID does not exist" containerID="9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.688168 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f"} err="failed to get container status \"9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f\": rpc error: code = NotFound desc = could not find container \"9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f\": container with ID starting with 9bbebaaf072a532f91f5b09a51f68f4108433b676b08c50cd20ccab70073508f not found: ID does not exist" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.777506 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "717cbf7e-4d14-4174-9de7-eba44c406e6e" (UID: "717cbf7e-4d14-4174-9de7-eba44c406e6e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.785708 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/717cbf7e-4d14-4174-9de7-eba44c406e6e-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.824858 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m4n54"] Sep 30 15:33:38 crc kubenswrapper[4799]: I0930 15:33:38.834779 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m4n54"] Sep 30 15:33:40 crc kubenswrapper[4799]: I0930 15:33:40.512971 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" path="/var/lib/kubelet/pods/717cbf7e-4d14-4174-9de7-eba44c406e6e/volumes" Sep 30 15:33:49 crc kubenswrapper[4799]: I0930 15:33:49.503717 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:33:49 crc kubenswrapper[4799]: E0930 15:33:49.504407 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:34:01 crc kubenswrapper[4799]: I0930 15:34:01.504041 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:34:01 crc kubenswrapper[4799]: E0930 15:34:01.505893 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:34:14 crc kubenswrapper[4799]: I0930 15:34:14.503212 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:34:14 crc kubenswrapper[4799]: E0930 15:34:14.504119 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:34:24 crc kubenswrapper[4799]: I0930 15:34:24.987113 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hzkbb"] Sep 30 15:34:24 crc kubenswrapper[4799]: E0930 15:34:24.988000 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerName="extract-content" Sep 30 15:34:24 crc kubenswrapper[4799]: I0930 15:34:24.988013 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerName="extract-content" Sep 30 15:34:24 crc kubenswrapper[4799]: E0930 15:34:24.988033 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerName="registry-server" Sep 30 15:34:24 crc kubenswrapper[4799]: I0930 15:34:24.988039 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerName="registry-server" Sep 30 15:34:24 crc kubenswrapper[4799]: E0930 15:34:24.988076 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerName="extract-utilities" Sep 30 15:34:24 crc kubenswrapper[4799]: I0930 15:34:24.988083 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerName="extract-utilities" Sep 30 15:34:24 crc kubenswrapper[4799]: I0930 15:34:24.988260 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="717cbf7e-4d14-4174-9de7-eba44c406e6e" containerName="registry-server" Sep 30 15:34:24 crc kubenswrapper[4799]: I0930 15:34:24.997506 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.008661 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hzkbb"] Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.142286 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5mbs\" (UniqueName: \"kubernetes.io/projected/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-kube-api-access-v5mbs\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.142355 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-catalog-content\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.142747 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-utilities\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.244238 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-utilities\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.244312 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5mbs\" (UniqueName: \"kubernetes.io/projected/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-kube-api-access-v5mbs\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.244349 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-catalog-content\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.244741 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-utilities\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.244988 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-catalog-content\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.280370 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5mbs\" (UniqueName: \"kubernetes.io/projected/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-kube-api-access-v5mbs\") pod \"community-operators-hzkbb\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:25 crc kubenswrapper[4799]: I0930 15:34:25.364773 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:26 crc kubenswrapper[4799]: I0930 15:34:26.796575 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hzkbb"] Sep 30 15:34:26 crc kubenswrapper[4799]: I0930 15:34:26.941553 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzkbb" event={"ID":"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f","Type":"ContainerStarted","Data":"3097cd62ddb30fbd413f8d5e966ce7777cb00c8009fdf2a7ecbad2666e45ead8"} Sep 30 15:34:27 crc kubenswrapper[4799]: I0930 15:34:27.952281 4799 generic.go:334] "Generic (PLEG): container finished" podID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerID="99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16" exitCode=0 Sep 30 15:34:27 crc kubenswrapper[4799]: I0930 15:34:27.952348 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzkbb" event={"ID":"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f","Type":"ContainerDied","Data":"99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16"} Sep 30 15:34:28 crc kubenswrapper[4799]: I0930 15:34:28.510323 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:34:28 crc kubenswrapper[4799]: E0930 15:34:28.510901 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:34:28 crc kubenswrapper[4799]: I0930 15:34:28.963258 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzkbb" event={"ID":"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f","Type":"ContainerStarted","Data":"44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09"} Sep 30 15:34:29 crc kubenswrapper[4799]: I0930 15:34:29.974537 4799 generic.go:334] "Generic (PLEG): container finished" podID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerID="44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09" exitCode=0 Sep 30 15:34:29 crc kubenswrapper[4799]: I0930 15:34:29.974634 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzkbb" event={"ID":"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f","Type":"ContainerDied","Data":"44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09"} Sep 30 15:34:30 crc kubenswrapper[4799]: I0930 15:34:30.987726 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzkbb" event={"ID":"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f","Type":"ContainerStarted","Data":"2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09"} Sep 30 15:34:31 crc kubenswrapper[4799]: I0930 15:34:31.011501 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hzkbb" podStartSLOduration=4.59554592 podStartE2EDuration="7.011478697s" podCreationTimestamp="2025-09-30 15:34:24 +0000 UTC" firstStartedPulling="2025-09-30 15:34:27.955405712 +0000 UTC m=+4490.039006129" lastFinishedPulling="2025-09-30 15:34:30.371338479 +0000 UTC m=+4492.454938906" observedRunningTime="2025-09-30 15:34:31.01056771 +0000 UTC m=+4493.094168157" watchObservedRunningTime="2025-09-30 15:34:31.011478697 +0000 UTC m=+4493.095079124" Sep 30 15:34:35 crc kubenswrapper[4799]: I0930 15:34:35.366116 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:35 crc kubenswrapper[4799]: I0930 15:34:35.366625 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:35 crc kubenswrapper[4799]: I0930 15:34:35.415279 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:36 crc kubenswrapper[4799]: I0930 15:34:36.082184 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:36 crc kubenswrapper[4799]: I0930 15:34:36.138079 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hzkbb"] Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.059942 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hzkbb" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerName="registry-server" containerID="cri-o://2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09" gracePeriod=2 Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.562145 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.719994 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-catalog-content\") pod \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.720063 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-utilities\") pod \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.720092 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5mbs\" (UniqueName: \"kubernetes.io/projected/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-kube-api-access-v5mbs\") pod \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\" (UID: \"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f\") " Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.721696 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-utilities" (OuterVolumeSpecName: "utilities") pod "bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" (UID: "bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.733743 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-kube-api-access-v5mbs" (OuterVolumeSpecName: "kube-api-access-v5mbs") pod "bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" (UID: "bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f"). InnerVolumeSpecName "kube-api-access-v5mbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.776316 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" (UID: "bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.822930 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.823004 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5mbs\" (UniqueName: \"kubernetes.io/projected/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-kube-api-access-v5mbs\") on node \"crc\" DevicePath \"\"" Sep 30 15:34:38 crc kubenswrapper[4799]: I0930 15:34:38.823023 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.071381 4799 generic.go:334] "Generic (PLEG): container finished" podID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerID="2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09" exitCode=0 Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.071435 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzkbb" event={"ID":"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f","Type":"ContainerDied","Data":"2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09"} Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.071442 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzkbb" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.071474 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzkbb" event={"ID":"bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f","Type":"ContainerDied","Data":"3097cd62ddb30fbd413f8d5e966ce7777cb00c8009fdf2a7ecbad2666e45ead8"} Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.071494 4799 scope.go:117] "RemoveContainer" containerID="2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.109782 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hzkbb"] Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.109994 4799 scope.go:117] "RemoveContainer" containerID="44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.120923 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hzkbb"] Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.147300 4799 scope.go:117] "RemoveContainer" containerID="99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.186590 4799 scope.go:117] "RemoveContainer" containerID="2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09" Sep 30 15:34:39 crc kubenswrapper[4799]: E0930 15:34:39.187136 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09\": container with ID starting with 2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09 not found: ID does not exist" containerID="2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.187179 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09"} err="failed to get container status \"2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09\": rpc error: code = NotFound desc = could not find container \"2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09\": container with ID starting with 2d16e37b44f7ad8328ead247929969cbffbaf2816543ca58c2a44280912bbd09 not found: ID does not exist" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.187206 4799 scope.go:117] "RemoveContainer" containerID="44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09" Sep 30 15:34:39 crc kubenswrapper[4799]: E0930 15:34:39.187686 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09\": container with ID starting with 44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09 not found: ID does not exist" containerID="44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.187743 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09"} err="failed to get container status \"44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09\": rpc error: code = NotFound desc = could not find container \"44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09\": container with ID starting with 44c37a0b8839a48d86f47cf9115f595aa5957bc23d29f32a95b4b65b258b6c09 not found: ID does not exist" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.187782 4799 scope.go:117] "RemoveContainer" containerID="99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16" Sep 30 15:34:39 crc kubenswrapper[4799]: E0930 15:34:39.188181 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16\": container with ID starting with 99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16 not found: ID does not exist" containerID="99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16" Sep 30 15:34:39 crc kubenswrapper[4799]: I0930 15:34:39.188215 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16"} err="failed to get container status \"99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16\": rpc error: code = NotFound desc = could not find container \"99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16\": container with ID starting with 99733580eb1954a5cab650d6339c9ec98c79333d0bda77af05f2822e5c4fed16 not found: ID does not exist" Sep 30 15:34:40 crc kubenswrapper[4799]: I0930 15:34:40.516057 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" path="/var/lib/kubelet/pods/bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f/volumes" Sep 30 15:34:43 crc kubenswrapper[4799]: I0930 15:34:43.503115 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:34:43 crc kubenswrapper[4799]: E0930 15:34:43.504151 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:34:57 crc kubenswrapper[4799]: I0930 15:34:57.503596 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:34:57 crc kubenswrapper[4799]: E0930 15:34:57.504495 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:35:11 crc kubenswrapper[4799]: I0930 15:35:11.503993 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:35:11 crc kubenswrapper[4799]: E0930 15:35:11.506430 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:35:25 crc kubenswrapper[4799]: I0930 15:35:25.503203 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:35:25 crc kubenswrapper[4799]: E0930 15:35:25.504100 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:35:40 crc kubenswrapper[4799]: I0930 15:35:40.503292 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:35:40 crc kubenswrapper[4799]: E0930 15:35:40.504158 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:35:55 crc kubenswrapper[4799]: I0930 15:35:55.504026 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:35:55 crc kubenswrapper[4799]: E0930 15:35:55.504986 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:36:08 crc kubenswrapper[4799]: I0930 15:36:08.510853 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:36:08 crc kubenswrapper[4799]: E0930 15:36:08.511600 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:36:19 crc kubenswrapper[4799]: I0930 15:36:19.503819 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:36:19 crc kubenswrapper[4799]: E0930 15:36:19.504634 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:36:31 crc kubenswrapper[4799]: I0930 15:36:31.504023 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:36:31 crc kubenswrapper[4799]: E0930 15:36:31.504937 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:36:44 crc kubenswrapper[4799]: I0930 15:36:44.503585 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:36:44 crc kubenswrapper[4799]: E0930 15:36:44.504476 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:36:58 crc kubenswrapper[4799]: I0930 15:36:58.509424 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:36:58 crc kubenswrapper[4799]: E0930 15:36:58.510249 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:37:12 crc kubenswrapper[4799]: I0930 15:37:12.505528 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:37:12 crc kubenswrapper[4799]: E0930 15:37:12.506401 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:37:26 crc kubenswrapper[4799]: I0930 15:37:26.503128 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:37:26 crc kubenswrapper[4799]: E0930 15:37:26.503881 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.430033 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9t8mm"] Sep 30 15:37:41 crc kubenswrapper[4799]: E0930 15:37:41.438916 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerName="extract-content" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.438947 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerName="extract-content" Sep 30 15:37:41 crc kubenswrapper[4799]: E0930 15:37:41.438956 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerName="extract-utilities" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.438965 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerName="extract-utilities" Sep 30 15:37:41 crc kubenswrapper[4799]: E0930 15:37:41.439012 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerName="registry-server" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.439020 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerName="registry-server" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.439288 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfa5d4dc-b5b8-49c7-a5f5-4802f991fd0f" containerName="registry-server" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.443460 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9t8mm"] Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.443613 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.472768 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbzs4\" (UniqueName: \"kubernetes.io/projected/451cc529-b354-4b50-b42e-4033e9d8deeb-kube-api-access-jbzs4\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.472954 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-catalog-content\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.473003 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-utilities\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.506520 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.579217 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbzs4\" (UniqueName: \"kubernetes.io/projected/451cc529-b354-4b50-b42e-4033e9d8deeb-kube-api-access-jbzs4\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.579903 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-catalog-content\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.579965 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-utilities\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.580526 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-utilities\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.581272 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-catalog-content\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.603429 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbzs4\" (UniqueName: \"kubernetes.io/projected/451cc529-b354-4b50-b42e-4033e9d8deeb-kube-api-access-jbzs4\") pod \"redhat-operators-9t8mm\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:41 crc kubenswrapper[4799]: I0930 15:37:41.777724 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:42 crc kubenswrapper[4799]: I0930 15:37:42.345261 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9t8mm"] Sep 30 15:37:42 crc kubenswrapper[4799]: I0930 15:37:42.739061 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"e4b97db586cc9e0bbe8e063b258329d9e848a6cab8be98493164f05db56b329d"} Sep 30 15:37:42 crc kubenswrapper[4799]: I0930 15:37:42.766203 4799 generic.go:334] "Generic (PLEG): container finished" podID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerID="aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e" exitCode=0 Sep 30 15:37:42 crc kubenswrapper[4799]: I0930 15:37:42.766269 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9t8mm" event={"ID":"451cc529-b354-4b50-b42e-4033e9d8deeb","Type":"ContainerDied","Data":"aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e"} Sep 30 15:37:42 crc kubenswrapper[4799]: I0930 15:37:42.766306 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9t8mm" event={"ID":"451cc529-b354-4b50-b42e-4033e9d8deeb","Type":"ContainerStarted","Data":"ce974fafdf4d825be9acd8c7a8c8329b54ddda05e939c093a249d9bbb68893f4"} Sep 30 15:37:44 crc kubenswrapper[4799]: I0930 15:37:44.784548 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9t8mm" event={"ID":"451cc529-b354-4b50-b42e-4033e9d8deeb","Type":"ContainerStarted","Data":"d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5"} Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.784062 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-stt42"] Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.787178 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.809563 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-stt42"] Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.838204 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db8pz\" (UniqueName: \"kubernetes.io/projected/cfb584e3-1f7d-41f4-9b68-298f3c609163-kube-api-access-db8pz\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.838311 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-utilities\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.838351 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-catalog-content\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.940580 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db8pz\" (UniqueName: \"kubernetes.io/projected/cfb584e3-1f7d-41f4-9b68-298f3c609163-kube-api-access-db8pz\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.940810 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-utilities\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.940890 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-catalog-content\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.941434 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-catalog-content\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.941427 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-utilities\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:46 crc kubenswrapper[4799]: I0930 15:37:46.979333 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db8pz\" (UniqueName: \"kubernetes.io/projected/cfb584e3-1f7d-41f4-9b68-298f3c609163-kube-api-access-db8pz\") pod \"redhat-marketplace-stt42\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:47 crc kubenswrapper[4799]: I0930 15:37:47.114702 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:47 crc kubenswrapper[4799]: I0930 15:37:47.617101 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-stt42"] Sep 30 15:37:47 crc kubenswrapper[4799]: I0930 15:37:47.817009 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stt42" event={"ID":"cfb584e3-1f7d-41f4-9b68-298f3c609163","Type":"ContainerStarted","Data":"e807916a35c3d55880f60c5419353c76180197c11704f39d9419b965e92b716f"} Sep 30 15:37:48 crc kubenswrapper[4799]: I0930 15:37:48.827058 4799 generic.go:334] "Generic (PLEG): container finished" podID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerID="143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c" exitCode=0 Sep 30 15:37:48 crc kubenswrapper[4799]: I0930 15:37:48.827191 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stt42" event={"ID":"cfb584e3-1f7d-41f4-9b68-298f3c609163","Type":"ContainerDied","Data":"143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c"} Sep 30 15:37:48 crc kubenswrapper[4799]: I0930 15:37:48.831727 4799 generic.go:334] "Generic (PLEG): container finished" podID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerID="d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5" exitCode=0 Sep 30 15:37:48 crc kubenswrapper[4799]: I0930 15:37:48.831775 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9t8mm" event={"ID":"451cc529-b354-4b50-b42e-4033e9d8deeb","Type":"ContainerDied","Data":"d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5"} Sep 30 15:37:50 crc kubenswrapper[4799]: I0930 15:37:50.854973 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9t8mm" event={"ID":"451cc529-b354-4b50-b42e-4033e9d8deeb","Type":"ContainerStarted","Data":"e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813"} Sep 30 15:37:50 crc kubenswrapper[4799]: I0930 15:37:50.887471 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9t8mm" podStartSLOduration=2.444581684 podStartE2EDuration="9.887447038s" podCreationTimestamp="2025-09-30 15:37:41 +0000 UTC" firstStartedPulling="2025-09-30 15:37:42.772494319 +0000 UTC m=+4684.856094746" lastFinishedPulling="2025-09-30 15:37:50.215359673 +0000 UTC m=+4692.298960100" observedRunningTime="2025-09-30 15:37:50.880288198 +0000 UTC m=+4692.963888635" watchObservedRunningTime="2025-09-30 15:37:50.887447038 +0000 UTC m=+4692.971047465" Sep 30 15:37:51 crc kubenswrapper[4799]: I0930 15:37:51.778866 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:51 crc kubenswrapper[4799]: I0930 15:37:51.779166 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:37:51 crc kubenswrapper[4799]: I0930 15:37:51.903186 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stt42" event={"ID":"cfb584e3-1f7d-41f4-9b68-298f3c609163","Type":"ContainerStarted","Data":"940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2"} Sep 30 15:37:52 crc kubenswrapper[4799]: I0930 15:37:52.823370 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9t8mm" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="registry-server" probeResult="failure" output=< Sep 30 15:37:52 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:37:52 crc kubenswrapper[4799]: > Sep 30 15:37:53 crc kubenswrapper[4799]: I0930 15:37:53.922163 4799 generic.go:334] "Generic (PLEG): container finished" podID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerID="940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2" exitCode=0 Sep 30 15:37:53 crc kubenswrapper[4799]: I0930 15:37:53.922271 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stt42" event={"ID":"cfb584e3-1f7d-41f4-9b68-298f3c609163","Type":"ContainerDied","Data":"940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2"} Sep 30 15:37:54 crc kubenswrapper[4799]: I0930 15:37:54.933728 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stt42" event={"ID":"cfb584e3-1f7d-41f4-9b68-298f3c609163","Type":"ContainerStarted","Data":"c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d"} Sep 30 15:37:54 crc kubenswrapper[4799]: I0930 15:37:54.964347 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-stt42" podStartSLOduration=3.3765937409999998 podStartE2EDuration="8.964324784s" podCreationTimestamp="2025-09-30 15:37:46 +0000 UTC" firstStartedPulling="2025-09-30 15:37:48.829489345 +0000 UTC m=+4690.913089772" lastFinishedPulling="2025-09-30 15:37:54.417220388 +0000 UTC m=+4696.500820815" observedRunningTime="2025-09-30 15:37:54.964209831 +0000 UTC m=+4697.047810278" watchObservedRunningTime="2025-09-30 15:37:54.964324784 +0000 UTC m=+4697.047925211" Sep 30 15:37:57 crc kubenswrapper[4799]: I0930 15:37:57.115536 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:57 crc kubenswrapper[4799]: I0930 15:37:57.115873 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:37:58 crc kubenswrapper[4799]: I0930 15:37:58.523413 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-stt42" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="registry-server" probeResult="failure" output=< Sep 30 15:37:58 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:37:58 crc kubenswrapper[4799]: > Sep 30 15:38:02 crc kubenswrapper[4799]: I0930 15:38:02.830714 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9t8mm" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="registry-server" probeResult="failure" output=< Sep 30 15:38:02 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:38:02 crc kubenswrapper[4799]: > Sep 30 15:38:04 crc kubenswrapper[4799]: I0930 15:38:04.017452 4799 generic.go:334] "Generic (PLEG): container finished" podID="ab210e62-906f-46ad-b80a-fbf00a875124" containerID="b8c2ed383ec607aa5c232aa30bcce5d5ad65e2ca9a50987192da6ae49739f4ab" exitCode=0 Sep 30 15:38:04 crc kubenswrapper[4799]: I0930 15:38:04.017579 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ab210e62-906f-46ad-b80a-fbf00a875124","Type":"ContainerDied","Data":"b8c2ed383ec607aa5c232aa30bcce5d5ad65e2ca9a50987192da6ae49739f4ab"} Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.585229 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642323 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz4m8\" (UniqueName: \"kubernetes.io/projected/ab210e62-906f-46ad-b80a-fbf00a875124-kube-api-access-nz4m8\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642405 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config-secret\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642444 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-workdir\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642490 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ssh-key\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642522 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-temporary\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642550 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-config-data\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642586 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642604 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.642739 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ca-certs\") pod \"ab210e62-906f-46ad-b80a-fbf00a875124\" (UID: \"ab210e62-906f-46ad-b80a-fbf00a875124\") " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.648866 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab210e62-906f-46ad-b80a-fbf00a875124-kube-api-access-nz4m8" (OuterVolumeSpecName: "kube-api-access-nz4m8") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "kube-api-access-nz4m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.658106 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.664397 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.677720 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.692703 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.692736 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-config-data" (OuterVolumeSpecName: "config-data") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.694603 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.696430 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.744761 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz4m8\" (UniqueName: \"kubernetes.io/projected/ab210e62-906f-46ad-b80a-fbf00a875124-kube-api-access-nz4m8\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.744799 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.744813 4799 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.744829 4799 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ssh-key\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.744841 4799 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ab210e62-906f-46ad-b80a-fbf00a875124-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.744854 4799 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-config-data\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.745412 4799 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.745440 4799 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ab210e62-906f-46ad-b80a-fbf00a875124-ca-certs\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.746939 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ab210e62-906f-46ad-b80a-fbf00a875124" (UID: "ab210e62-906f-46ad-b80a-fbf00a875124"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.774072 4799 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.846931 4799 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:05 crc kubenswrapper[4799]: I0930 15:38:05.846967 4799 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ab210e62-906f-46ad-b80a-fbf00a875124-openstack-config\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:06 crc kubenswrapper[4799]: I0930 15:38:06.044814 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ab210e62-906f-46ad-b80a-fbf00a875124","Type":"ContainerDied","Data":"a8a4f366047149510b6bd1463e01c76a42804713ad28e9300ce24a04b64f8d89"} Sep 30 15:38:06 crc kubenswrapper[4799]: I0930 15:38:06.044870 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8a4f366047149510b6bd1463e01c76a42804713ad28e9300ce24a04b64f8d89" Sep 30 15:38:06 crc kubenswrapper[4799]: I0930 15:38:06.044962 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Sep 30 15:38:07 crc kubenswrapper[4799]: I0930 15:38:07.170766 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:38:07 crc kubenswrapper[4799]: I0930 15:38:07.224341 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:38:07 crc kubenswrapper[4799]: I0930 15:38:07.415032 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-stt42"] Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.074217 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-stt42" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="registry-server" containerID="cri-o://c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d" gracePeriod=2 Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.581356 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.625966 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-catalog-content\") pod \"cfb584e3-1f7d-41f4-9b68-298f3c609163\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.626380 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db8pz\" (UniqueName: \"kubernetes.io/projected/cfb584e3-1f7d-41f4-9b68-298f3c609163-kube-api-access-db8pz\") pod \"cfb584e3-1f7d-41f4-9b68-298f3c609163\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.626468 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-utilities\") pod \"cfb584e3-1f7d-41f4-9b68-298f3c609163\" (UID: \"cfb584e3-1f7d-41f4-9b68-298f3c609163\") " Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.629050 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-utilities" (OuterVolumeSpecName: "utilities") pod "cfb584e3-1f7d-41f4-9b68-298f3c609163" (UID: "cfb584e3-1f7d-41f4-9b68-298f3c609163"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.650201 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfb584e3-1f7d-41f4-9b68-298f3c609163" (UID: "cfb584e3-1f7d-41f4-9b68-298f3c609163"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.661863 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb584e3-1f7d-41f4-9b68-298f3c609163-kube-api-access-db8pz" (OuterVolumeSpecName: "kube-api-access-db8pz") pod "cfb584e3-1f7d-41f4-9b68-298f3c609163" (UID: "cfb584e3-1f7d-41f4-9b68-298f3c609163"). InnerVolumeSpecName "kube-api-access-db8pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.729237 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db8pz\" (UniqueName: \"kubernetes.io/projected/cfb584e3-1f7d-41f4-9b68-298f3c609163-kube-api-access-db8pz\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.729562 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:09 crc kubenswrapper[4799]: I0930 15:38:09.729576 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb584e3-1f7d-41f4-9b68-298f3c609163-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.085419 4799 generic.go:334] "Generic (PLEG): container finished" podID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerID="c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d" exitCode=0 Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.085473 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stt42" event={"ID":"cfb584e3-1f7d-41f4-9b68-298f3c609163","Type":"ContainerDied","Data":"c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d"} Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.085539 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stt42" event={"ID":"cfb584e3-1f7d-41f4-9b68-298f3c609163","Type":"ContainerDied","Data":"e807916a35c3d55880f60c5419353c76180197c11704f39d9419b965e92b716f"} Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.085536 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-stt42" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.085560 4799 scope.go:117] "RemoveContainer" containerID="c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.114246 4799 scope.go:117] "RemoveContainer" containerID="940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.133549 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-stt42"] Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.140575 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-stt42"] Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.390010 4799 scope.go:117] "RemoveContainer" containerID="143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.460559 4799 scope.go:117] "RemoveContainer" containerID="c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d" Sep 30 15:38:10 crc kubenswrapper[4799]: E0930 15:38:10.461924 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d\": container with ID starting with c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d not found: ID does not exist" containerID="c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.462008 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d"} err="failed to get container status \"c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d\": rpc error: code = NotFound desc = could not find container \"c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d\": container with ID starting with c52b8ce8f62e092969512950d1acb0b9878344ac792bbc18b60ec8ce3fbfed0d not found: ID does not exist" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.462071 4799 scope.go:117] "RemoveContainer" containerID="940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2" Sep 30 15:38:10 crc kubenswrapper[4799]: E0930 15:38:10.463507 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2\": container with ID starting with 940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2 not found: ID does not exist" containerID="940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.463546 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2"} err="failed to get container status \"940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2\": rpc error: code = NotFound desc = could not find container \"940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2\": container with ID starting with 940646c7f6976c6952c2a62b169d2719612edcdd015c421fd40d5ddfbb7fb9c2 not found: ID does not exist" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.463630 4799 scope.go:117] "RemoveContainer" containerID="143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c" Sep 30 15:38:10 crc kubenswrapper[4799]: E0930 15:38:10.465013 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c\": container with ID starting with 143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c not found: ID does not exist" containerID="143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.465043 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c"} err="failed to get container status \"143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c\": rpc error: code = NotFound desc = could not find container \"143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c\": container with ID starting with 143ad0aab7e0b29be1c06e9ed5108da5fa0de48a7f690c50b67ce22a04abf69c not found: ID does not exist" Sep 30 15:38:10 crc kubenswrapper[4799]: I0930 15:38:10.515928 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" path="/var/lib/kubelet/pods/cfb584e3-1f7d-41f4-9b68-298f3c609163/volumes" Sep 30 15:38:12 crc kubenswrapper[4799]: I0930 15:38:12.318470 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:38:12 crc kubenswrapper[4799]: I0930 15:38:12.374491 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:38:12 crc kubenswrapper[4799]: I0930 15:38:12.816581 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9t8mm"] Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.125841 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9t8mm" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="registry-server" containerID="cri-o://e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813" gracePeriod=2 Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.608148 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 15:38:14 crc kubenswrapper[4799]: E0930 15:38:14.609273 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab210e62-906f-46ad-b80a-fbf00a875124" containerName="tempest-tests-tempest-tests-runner" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.609310 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab210e62-906f-46ad-b80a-fbf00a875124" containerName="tempest-tests-tempest-tests-runner" Sep 30 15:38:14 crc kubenswrapper[4799]: E0930 15:38:14.609339 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="registry-server" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.609348 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="registry-server" Sep 30 15:38:14 crc kubenswrapper[4799]: E0930 15:38:14.609369 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="extract-utilities" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.609376 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="extract-utilities" Sep 30 15:38:14 crc kubenswrapper[4799]: E0930 15:38:14.609410 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="extract-content" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.609417 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="extract-content" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.609669 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb584e3-1f7d-41f4-9b68-298f3c609163" containerName="registry-server" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.609707 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab210e62-906f-46ad-b80a-fbf00a875124" containerName="tempest-tests-tempest-tests-runner" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.610481 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.619035 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5f8xr" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.626135 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.757644 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.764030 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b3f82270-84d2-409f-b243-cde51bfb251f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.764203 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmjjf\" (UniqueName: \"kubernetes.io/projected/b3f82270-84d2-409f-b243-cde51bfb251f-kube-api-access-mmjjf\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b3f82270-84d2-409f-b243-cde51bfb251f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.865562 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-catalog-content\") pod \"451cc529-b354-4b50-b42e-4033e9d8deeb\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.865702 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbzs4\" (UniqueName: \"kubernetes.io/projected/451cc529-b354-4b50-b42e-4033e9d8deeb-kube-api-access-jbzs4\") pod \"451cc529-b354-4b50-b42e-4033e9d8deeb\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.865840 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-utilities\") pod \"451cc529-b354-4b50-b42e-4033e9d8deeb\" (UID: \"451cc529-b354-4b50-b42e-4033e9d8deeb\") " Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.867453 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmjjf\" (UniqueName: \"kubernetes.io/projected/b3f82270-84d2-409f-b243-cde51bfb251f-kube-api-access-mmjjf\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b3f82270-84d2-409f-b243-cde51bfb251f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.867571 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b3f82270-84d2-409f-b243-cde51bfb251f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.867451 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-utilities" (OuterVolumeSpecName: "utilities") pod "451cc529-b354-4b50-b42e-4033e9d8deeb" (UID: "451cc529-b354-4b50-b42e-4033e9d8deeb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.868170 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.869957 4799 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b3f82270-84d2-409f-b243-cde51bfb251f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.871884 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/451cc529-b354-4b50-b42e-4033e9d8deeb-kube-api-access-jbzs4" (OuterVolumeSpecName: "kube-api-access-jbzs4") pod "451cc529-b354-4b50-b42e-4033e9d8deeb" (UID: "451cc529-b354-4b50-b42e-4033e9d8deeb"). InnerVolumeSpecName "kube-api-access-jbzs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.893024 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmjjf\" (UniqueName: \"kubernetes.io/projected/b3f82270-84d2-409f-b243-cde51bfb251f-kube-api-access-mmjjf\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b3f82270-84d2-409f-b243-cde51bfb251f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.901559 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b3f82270-84d2-409f-b243-cde51bfb251f\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.955522 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "451cc529-b354-4b50-b42e-4033e9d8deeb" (UID: "451cc529-b354-4b50-b42e-4033e9d8deeb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.969951 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451cc529-b354-4b50-b42e-4033e9d8deeb-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:14 crc kubenswrapper[4799]: I0930 15:38:14.970001 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbzs4\" (UniqueName: \"kubernetes.io/projected/451cc529-b354-4b50-b42e-4033e9d8deeb-kube-api-access-jbzs4\") on node \"crc\" DevicePath \"\"" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.053165 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.138812 4799 generic.go:334] "Generic (PLEG): container finished" podID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerID="e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813" exitCode=0 Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.138873 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9t8mm" event={"ID":"451cc529-b354-4b50-b42e-4033e9d8deeb","Type":"ContainerDied","Data":"e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813"} Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.138907 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9t8mm" event={"ID":"451cc529-b354-4b50-b42e-4033e9d8deeb","Type":"ContainerDied","Data":"ce974fafdf4d825be9acd8c7a8c8329b54ddda05e939c093a249d9bbb68893f4"} Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.138925 4799 scope.go:117] "RemoveContainer" containerID="e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.139079 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9t8mm" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.211071 4799 scope.go:117] "RemoveContainer" containerID="d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.211756 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9t8mm"] Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.224515 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9t8mm"] Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.259865 4799 scope.go:117] "RemoveContainer" containerID="aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.356260 4799 scope.go:117] "RemoveContainer" containerID="e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813" Sep 30 15:38:15 crc kubenswrapper[4799]: E0930 15:38:15.365340 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813\": container with ID starting with e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813 not found: ID does not exist" containerID="e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.365395 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813"} err="failed to get container status \"e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813\": rpc error: code = NotFound desc = could not find container \"e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813\": container with ID starting with e14927608d7c41d0619de5623beb50d38ccc53db0573b42d20877573e1ea6813 not found: ID does not exist" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.365435 4799 scope.go:117] "RemoveContainer" containerID="d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5" Sep 30 15:38:15 crc kubenswrapper[4799]: E0930 15:38:15.366141 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5\": container with ID starting with d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5 not found: ID does not exist" containerID="d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.366170 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5"} err="failed to get container status \"d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5\": rpc error: code = NotFound desc = could not find container \"d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5\": container with ID starting with d1570becfe301dcd7a108c8048822bdfd340e597181038876bb6d92efd3565c5 not found: ID does not exist" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.366189 4799 scope.go:117] "RemoveContainer" containerID="aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e" Sep 30 15:38:15 crc kubenswrapper[4799]: E0930 15:38:15.366465 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e\": container with ID starting with aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e not found: ID does not exist" containerID="aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.366491 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e"} err="failed to get container status \"aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e\": rpc error: code = NotFound desc = could not find container \"aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e\": container with ID starting with aed7a6db516a63dce5b1264d0098da95411eaea60e92f3b5f7b2b44a80f5b02e not found: ID does not exist" Sep 30 15:38:15 crc kubenswrapper[4799]: I0930 15:38:15.603828 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Sep 30 15:38:16 crc kubenswrapper[4799]: I0930 15:38:16.161306 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b3f82270-84d2-409f-b243-cde51bfb251f","Type":"ContainerStarted","Data":"4180048e8b1310de9698f600b3be5e61b4f0550e099cb4e94ac06620d33f8b01"} Sep 30 15:38:16 crc kubenswrapper[4799]: I0930 15:38:16.514342 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" path="/var/lib/kubelet/pods/451cc529-b354-4b50-b42e-4033e9d8deeb/volumes" Sep 30 15:38:17 crc kubenswrapper[4799]: I0930 15:38:17.171101 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b3f82270-84d2-409f-b243-cde51bfb251f","Type":"ContainerStarted","Data":"fce71a254352bb5ea2f38f2da232ecec1e371a45edf6ea560405487e634d551b"} Sep 30 15:38:17 crc kubenswrapper[4799]: I0930 15:38:17.190020 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.241651659 podStartE2EDuration="3.189968291s" podCreationTimestamp="2025-09-30 15:38:14 +0000 UTC" firstStartedPulling="2025-09-30 15:38:15.61395393 +0000 UTC m=+4717.697554357" lastFinishedPulling="2025-09-30 15:38:16.562270562 +0000 UTC m=+4718.645870989" observedRunningTime="2025-09-30 15:38:17.184042798 +0000 UTC m=+4719.267643235" watchObservedRunningTime="2025-09-30 15:38:17.189968291 +0000 UTC m=+4719.273568758" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.827438 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d2c2l/must-gather-njlvq"] Sep 30 15:38:34 crc kubenswrapper[4799]: E0930 15:38:34.828436 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="extract-content" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.828453 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="extract-content" Sep 30 15:38:34 crc kubenswrapper[4799]: E0930 15:38:34.828475 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="extract-utilities" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.828481 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="extract-utilities" Sep 30 15:38:34 crc kubenswrapper[4799]: E0930 15:38:34.828500 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="registry-server" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.828508 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="registry-server" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.828717 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="451cc529-b354-4b50-b42e-4033e9d8deeb" containerName="registry-server" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.829905 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.832302 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-d2c2l"/"openshift-service-ca.crt" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.832302 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-d2c2l"/"kube-root-ca.crt" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.835145 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-d2c2l"/"default-dockercfg-7tvdx" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.853787 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-d2c2l/must-gather-njlvq"] Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.959302 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cad0aa86-fab9-465c-9062-00d77763387b-must-gather-output\") pod \"must-gather-njlvq\" (UID: \"cad0aa86-fab9-465c-9062-00d77763387b\") " pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:38:34 crc kubenswrapper[4799]: I0930 15:38:34.959386 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm5kq\" (UniqueName: \"kubernetes.io/projected/cad0aa86-fab9-465c-9062-00d77763387b-kube-api-access-wm5kq\") pod \"must-gather-njlvq\" (UID: \"cad0aa86-fab9-465c-9062-00d77763387b\") " pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:38:35 crc kubenswrapper[4799]: I0930 15:38:35.061004 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cad0aa86-fab9-465c-9062-00d77763387b-must-gather-output\") pod \"must-gather-njlvq\" (UID: \"cad0aa86-fab9-465c-9062-00d77763387b\") " pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:38:35 crc kubenswrapper[4799]: I0930 15:38:35.061318 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm5kq\" (UniqueName: \"kubernetes.io/projected/cad0aa86-fab9-465c-9062-00d77763387b-kube-api-access-wm5kq\") pod \"must-gather-njlvq\" (UID: \"cad0aa86-fab9-465c-9062-00d77763387b\") " pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:38:35 crc kubenswrapper[4799]: I0930 15:38:35.061508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cad0aa86-fab9-465c-9062-00d77763387b-must-gather-output\") pod \"must-gather-njlvq\" (UID: \"cad0aa86-fab9-465c-9062-00d77763387b\") " pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:38:35 crc kubenswrapper[4799]: I0930 15:38:35.085182 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm5kq\" (UniqueName: \"kubernetes.io/projected/cad0aa86-fab9-465c-9062-00d77763387b-kube-api-access-wm5kq\") pod \"must-gather-njlvq\" (UID: \"cad0aa86-fab9-465c-9062-00d77763387b\") " pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:38:35 crc kubenswrapper[4799]: I0930 15:38:35.151897 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:38:35 crc kubenswrapper[4799]: I0930 15:38:35.731308 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-d2c2l/must-gather-njlvq"] Sep 30 15:38:35 crc kubenswrapper[4799]: I0930 15:38:35.738672 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 15:38:36 crc kubenswrapper[4799]: I0930 15:38:36.423341 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/must-gather-njlvq" event={"ID":"cad0aa86-fab9-465c-9062-00d77763387b","Type":"ContainerStarted","Data":"92ff367438314bc87e8f1446a087a9e004c2751d0061849860a8bbf1d1f83bbe"} Sep 30 15:38:43 crc kubenswrapper[4799]: I0930 15:38:43.491867 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/must-gather-njlvq" event={"ID":"cad0aa86-fab9-465c-9062-00d77763387b","Type":"ContainerStarted","Data":"ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a"} Sep 30 15:38:43 crc kubenswrapper[4799]: I0930 15:38:43.493903 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/must-gather-njlvq" event={"ID":"cad0aa86-fab9-465c-9062-00d77763387b","Type":"ContainerStarted","Data":"8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80"} Sep 30 15:38:43 crc kubenswrapper[4799]: I0930 15:38:43.511222 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d2c2l/must-gather-njlvq" podStartSLOduration=2.49621046 podStartE2EDuration="9.51119605s" podCreationTimestamp="2025-09-30 15:38:34 +0000 UTC" firstStartedPulling="2025-09-30 15:38:35.738324619 +0000 UTC m=+4737.821925046" lastFinishedPulling="2025-09-30 15:38:42.753310209 +0000 UTC m=+4744.836910636" observedRunningTime="2025-09-30 15:38:43.507612756 +0000 UTC m=+4745.591213193" watchObservedRunningTime="2025-09-30 15:38:43.51119605 +0000 UTC m=+4745.594796477" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.095491 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-5tkj9"] Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.100785 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.196829 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9psmp\" (UniqueName: \"kubernetes.io/projected/88ef93f5-1f51-4863-bf88-57e75d415b68-kube-api-access-9psmp\") pod \"crc-debug-5tkj9\" (UID: \"88ef93f5-1f51-4863-bf88-57e75d415b68\") " pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.196948 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ef93f5-1f51-4863-bf88-57e75d415b68-host\") pod \"crc-debug-5tkj9\" (UID: \"88ef93f5-1f51-4863-bf88-57e75d415b68\") " pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.299334 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ef93f5-1f51-4863-bf88-57e75d415b68-host\") pod \"crc-debug-5tkj9\" (UID: \"88ef93f5-1f51-4863-bf88-57e75d415b68\") " pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.299463 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9psmp\" (UniqueName: \"kubernetes.io/projected/88ef93f5-1f51-4863-bf88-57e75d415b68-kube-api-access-9psmp\") pod \"crc-debug-5tkj9\" (UID: \"88ef93f5-1f51-4863-bf88-57e75d415b68\") " pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.299542 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ef93f5-1f51-4863-bf88-57e75d415b68-host\") pod \"crc-debug-5tkj9\" (UID: \"88ef93f5-1f51-4863-bf88-57e75d415b68\") " pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.340192 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9psmp\" (UniqueName: \"kubernetes.io/projected/88ef93f5-1f51-4863-bf88-57e75d415b68-kube-api-access-9psmp\") pod \"crc-debug-5tkj9\" (UID: \"88ef93f5-1f51-4863-bf88-57e75d415b68\") " pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.448909 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:38:50 crc kubenswrapper[4799]: I0930 15:38:50.572367 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" event={"ID":"88ef93f5-1f51-4863-bf88-57e75d415b68","Type":"ContainerStarted","Data":"ff96424c2a13565862df532e8c937f52a329f0fa5f53ccdc7342ccd7de04ac1d"} Sep 30 15:39:03 crc kubenswrapper[4799]: I0930 15:39:03.735648 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" event={"ID":"88ef93f5-1f51-4863-bf88-57e75d415b68","Type":"ContainerStarted","Data":"997a7fd3789451e743f96887058dd1ab8ea90dcb13c08271a8f88c8f0ef9e903"} Sep 30 15:39:03 crc kubenswrapper[4799]: I0930 15:39:03.758491 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" podStartSLOduration=1.6476168580000001 podStartE2EDuration="13.758471018s" podCreationTimestamp="2025-09-30 15:38:50 +0000 UTC" firstStartedPulling="2025-09-30 15:38:50.509404931 +0000 UTC m=+4752.593005368" lastFinishedPulling="2025-09-30 15:39:02.620259101 +0000 UTC m=+4764.703859528" observedRunningTime="2025-09-30 15:39:03.749588069 +0000 UTC m=+4765.833188496" watchObservedRunningTime="2025-09-30 15:39:03.758471018 +0000 UTC m=+4765.842071445" Sep 30 15:39:59 crc kubenswrapper[4799]: I0930 15:39:59.648517 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:39:59 crc kubenswrapper[4799]: I0930 15:39:59.649939 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:40:25 crc kubenswrapper[4799]: I0930 15:40:25.795911 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84f5689c94-tlpq6_b992c9d0-6ae3-4032-a4d5-f893c4813931/barbican-api/0.log" Sep 30 15:40:25 crc kubenswrapper[4799]: I0930 15:40:25.847660 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84f5689c94-tlpq6_b992c9d0-6ae3-4032-a4d5-f893c4813931/barbican-api-log/0.log" Sep 30 15:40:26 crc kubenswrapper[4799]: I0930 15:40:26.322134 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5888b5f87b-kmzgb_663d1769-2ce0-436c-986e-b07abcb30dca/barbican-keystone-listener/0.log" Sep 30 15:40:26 crc kubenswrapper[4799]: I0930 15:40:26.493563 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5888b5f87b-kmzgb_663d1769-2ce0-436c-986e-b07abcb30dca/barbican-keystone-listener-log/0.log" Sep 30 15:40:26 crc kubenswrapper[4799]: I0930 15:40:26.736830 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7958b68d9f-qk798_e704a0e9-df6f-46c9-a9fd-8e5547031469/barbican-worker/0.log" Sep 30 15:40:26 crc kubenswrapper[4799]: I0930 15:40:26.776110 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7958b68d9f-qk798_e704a0e9-df6f-46c9-a9fd-8e5547031469/barbican-worker-log/0.log" Sep 30 15:40:27 crc kubenswrapper[4799]: I0930 15:40:27.001500 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl_4565a6db-10d1-440b-a7b7-dc338c18f70f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:27 crc kubenswrapper[4799]: I0930 15:40:27.141239 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a/ceilometer-central-agent/0.log" Sep 30 15:40:27 crc kubenswrapper[4799]: I0930 15:40:27.281210 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a/ceilometer-notification-agent/0.log" Sep 30 15:40:27 crc kubenswrapper[4799]: I0930 15:40:27.342841 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a/proxy-httpd/0.log" Sep 30 15:40:27 crc kubenswrapper[4799]: I0930 15:40:27.441859 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a/sg-core/0.log" Sep 30 15:40:27 crc kubenswrapper[4799]: I0930 15:40:27.724693 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_da52c15f-1f9f-4485-99bc-1ea6b29c94c8/cinder-api/0.log" Sep 30 15:40:27 crc kubenswrapper[4799]: I0930 15:40:27.760040 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_da52c15f-1f9f-4485-99bc-1ea6b29c94c8/cinder-api-log/0.log" Sep 30 15:40:27 crc kubenswrapper[4799]: I0930 15:40:27.973853 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5c94a3cf-a75d-469a-9c13-ca9dd2ff978b/cinder-scheduler/0.log" Sep 30 15:40:28 crc kubenswrapper[4799]: I0930 15:40:28.120675 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5c94a3cf-a75d-469a-9c13-ca9dd2ff978b/probe/0.log" Sep 30 15:40:28 crc kubenswrapper[4799]: I0930 15:40:28.357384 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt_2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:28 crc kubenswrapper[4799]: I0930 15:40:28.966617 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-llf6t_e2cb3e85-dce2-4274-9472-3d5b10ba4358/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:29 crc kubenswrapper[4799]: I0930 15:40:29.239584 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-q5dph_200d7385-c387-4fc1-b108-fd1873d6da2d/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:29 crc kubenswrapper[4799]: I0930 15:40:29.284994 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-s2gr4_9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd/init/0.log" Sep 30 15:40:29 crc kubenswrapper[4799]: I0930 15:40:29.564297 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-s2gr4_9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd/init/0.log" Sep 30 15:40:29 crc kubenswrapper[4799]: I0930 15:40:29.648477 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:40:29 crc kubenswrapper[4799]: I0930 15:40:29.648543 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:40:29 crc kubenswrapper[4799]: I0930 15:40:29.857744 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-s2gr4_9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd/dnsmasq-dns/0.log" Sep 30 15:40:29 crc kubenswrapper[4799]: I0930 15:40:29.877709 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-svwrv_8da29a07-04e5-4c03-adf7-f642c86abecd/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:30 crc kubenswrapper[4799]: I0930 15:40:30.600340 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0334f39c-a7d6-42f3-a764-04e8e4857c41/glance-log/0.log" Sep 30 15:40:30 crc kubenswrapper[4799]: I0930 15:40:30.651444 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0334f39c-a7d6-42f3-a764-04e8e4857c41/glance-httpd/0.log" Sep 30 15:40:30 crc kubenswrapper[4799]: I0930 15:40:30.717237 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5bbb01bf-9925-4813-94e7-17882aa4d4c5/glance-httpd/0.log" Sep 30 15:40:30 crc kubenswrapper[4799]: I0930 15:40:30.848374 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5bbb01bf-9925-4813-94e7-17882aa4d4c5/glance-log/0.log" Sep 30 15:40:31 crc kubenswrapper[4799]: I0930 15:40:31.054563 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f7c888686-pwv2q_bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6/horizon/4.log" Sep 30 15:40:31 crc kubenswrapper[4799]: I0930 15:40:31.176631 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f7c888686-pwv2q_bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6/horizon/3.log" Sep 30 15:40:31 crc kubenswrapper[4799]: I0930 15:40:31.405111 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb_b410520b-9fd5-436d-9f2d-48a9ae6e517a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:31 crc kubenswrapper[4799]: I0930 15:40:31.730039 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6f7fz_9a20125e-0fe9-4826-b17d-f7b61c73bac3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:31 crc kubenswrapper[4799]: I0930 15:40:31.862300 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f7c888686-pwv2q_bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6/horizon-log/0.log" Sep 30 15:40:31 crc kubenswrapper[4799]: I0930 15:40:31.988178 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29320741-x298c_6f205373-c895-453e-90bd-9c323f6c90ac/keystone-cron/0.log" Sep 30 15:40:32 crc kubenswrapper[4799]: I0930 15:40:32.269107 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b4ecfc5d-9019-459f-83e8-8affcb145e78/kube-state-metrics/0.log" Sep 30 15:40:32 crc kubenswrapper[4799]: I0930 15:40:32.567117 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-577d747d48-ldwh6_301834e4-5636-4faa-b4b3-2b98f14791a7/keystone-api/0.log" Sep 30 15:40:32 crc kubenswrapper[4799]: I0930 15:40:32.636015 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw_43ae8f90-c778-47e6-8c69-8cdd2b4fa949/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:33 crc kubenswrapper[4799]: I0930 15:40:33.544576 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6_4ead1575-9ec6-41d9-8eda-f986d3a28713/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:33 crc kubenswrapper[4799]: I0930 15:40:33.587333 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-67c5b7d44f-24zhg_c013c008-ebe8-4afe-8a76-b9103ae99fa6/neutron-httpd/0.log" Sep 30 15:40:33 crc kubenswrapper[4799]: I0930 15:40:33.860959 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e3bd7a13-1fa3-4901-aeb6-3b9011ae330b/memcached/0.log" Sep 30 15:40:33 crc kubenswrapper[4799]: I0930 15:40:33.861375 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-67c5b7d44f-24zhg_c013c008-ebe8-4afe-8a76-b9103ae99fa6/neutron-api/0.log" Sep 30 15:40:34 crc kubenswrapper[4799]: I0930 15:40:34.620155 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_4f55c7c0-30d5-4349-956a-c06acef9fba0/nova-cell0-conductor-conductor/0.log" Sep 30 15:40:35 crc kubenswrapper[4799]: I0930 15:40:35.036131 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ca946c79-58c1-4f6b-9f61-e685bc6dcda9/nova-cell1-conductor-conductor/0.log" Sep 30 15:40:35 crc kubenswrapper[4799]: I0930 15:40:35.338231 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_a5fd7be8-ed88-4d3c-b979-8d944ca7b864/nova-cell1-novncproxy-novncproxy/0.log" Sep 30 15:40:35 crc kubenswrapper[4799]: I0930 15:40:35.400562 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_28bc13f2-5895-42cf-8ab5-60251eb4d09f/nova-api-log/0.log" Sep 30 15:40:35 crc kubenswrapper[4799]: I0930 15:40:35.444440 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_28bc13f2-5895-42cf-8ab5-60251eb4d09f/nova-api-api/0.log" Sep 30 15:40:35 crc kubenswrapper[4799]: I0930 15:40:35.653796 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-tlstx_2da7cba0-9bbc-47c7-99b0-f204928886fa/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:35 crc kubenswrapper[4799]: I0930 15:40:35.852741 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_fad4220f-11f3-45e3-b5f5-70d210ad5f3f/nova-metadata-log/0.log" Sep 30 15:40:36 crc kubenswrapper[4799]: I0930 15:40:36.363299 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_33a2e46f-eaef-4d54-8af4-ef388d3425dc/mysql-bootstrap/0.log" Sep 30 15:40:36 crc kubenswrapper[4799]: I0930 15:40:36.523012 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_886bc5b6-844e-4a31-8fe2-d62c56108e3a/nova-scheduler-scheduler/0.log" Sep 30 15:40:36 crc kubenswrapper[4799]: I0930 15:40:36.675071 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_33a2e46f-eaef-4d54-8af4-ef388d3425dc/mysql-bootstrap/0.log" Sep 30 15:40:36 crc kubenswrapper[4799]: I0930 15:40:36.683592 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_33a2e46f-eaef-4d54-8af4-ef388d3425dc/galera/0.log" Sep 30 15:40:37 crc kubenswrapper[4799]: I0930 15:40:37.039973 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b86436d8-96e2-44aa-9b4c-b9292a1ca129/mysql-bootstrap/0.log" Sep 30 15:40:37 crc kubenswrapper[4799]: I0930 15:40:37.172386 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_fad4220f-11f3-45e3-b5f5-70d210ad5f3f/nova-metadata-metadata/0.log" Sep 30 15:40:37 crc kubenswrapper[4799]: I0930 15:40:37.300833 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b86436d8-96e2-44aa-9b4c-b9292a1ca129/mysql-bootstrap/0.log" Sep 30 15:40:37 crc kubenswrapper[4799]: I0930 15:40:37.337108 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b86436d8-96e2-44aa-9b4c-b9292a1ca129/galera/0.log" Sep 30 15:40:37 crc kubenswrapper[4799]: I0930 15:40:37.410734 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_93dd541f-bdad-44fb-9450-1baa7acabb40/openstackclient/0.log" Sep 30 15:40:37 crc kubenswrapper[4799]: I0930 15:40:37.579598 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2kvml_cbefa245-5ad5-4303-b166-59574e8a0c65/ovn-controller/0.log" Sep 30 15:40:37 crc kubenswrapper[4799]: I0930 15:40:37.670427 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-r82xm_9ee89b02-9e5e-4c62-ade3-c4a308feefde/openstack-network-exporter/0.log" Sep 30 15:40:37 crc kubenswrapper[4799]: I0930 15:40:37.784195 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v2rs4_17bda665-adad-40d3-9465-46013020af4e/ovsdb-server-init/0.log" Sep 30 15:40:38 crc kubenswrapper[4799]: I0930 15:40:38.005601 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v2rs4_17bda665-adad-40d3-9465-46013020af4e/ovsdb-server-init/0.log" Sep 30 15:40:38 crc kubenswrapper[4799]: I0930 15:40:38.048682 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v2rs4_17bda665-adad-40d3-9465-46013020af4e/ovs-vswitchd/0.log" Sep 30 15:40:38 crc kubenswrapper[4799]: I0930 15:40:38.091616 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v2rs4_17bda665-adad-40d3-9465-46013020af4e/ovsdb-server/0.log" Sep 30 15:40:38 crc kubenswrapper[4799]: I0930 15:40:38.579929 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_583080ea-2fdf-461a-8b59-f04c48054eb9/openstack-network-exporter/0.log" Sep 30 15:40:38 crc kubenswrapper[4799]: I0930 15:40:38.597491 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-499wk_7110fd32-977a-4666-8dc1-d0d217767662/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:38 crc kubenswrapper[4799]: I0930 15:40:38.879114 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c146ba83-511e-4ff2-9a3e-6c83c4c43e9e/openstack-network-exporter/0.log" Sep 30 15:40:38 crc kubenswrapper[4799]: I0930 15:40:38.882361 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_583080ea-2fdf-461a-8b59-f04c48054eb9/ovn-northd/0.log" Sep 30 15:40:38 crc kubenswrapper[4799]: I0930 15:40:38.968155 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c146ba83-511e-4ff2-9a3e-6c83c4c43e9e/ovsdbserver-nb/0.log" Sep 30 15:40:39 crc kubenswrapper[4799]: I0930 15:40:39.092351 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8d8fe29c-5270-400a-9fc6-1024482dd54d/openstack-network-exporter/0.log" Sep 30 15:40:39 crc kubenswrapper[4799]: I0930 15:40:39.190502 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8d8fe29c-5270-400a-9fc6-1024482dd54d/ovsdbserver-sb/0.log" Sep 30 15:40:39 crc kubenswrapper[4799]: I0930 15:40:39.429430 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9fd6b988d-fckfn_23d1089b-5d01-44ec-b8de-617485c30c7a/placement-api/0.log" Sep 30 15:40:39 crc kubenswrapper[4799]: I0930 15:40:39.577927 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9fd6b988d-fckfn_23d1089b-5d01-44ec-b8de-617485c30c7a/placement-log/0.log" Sep 30 15:40:39 crc kubenswrapper[4799]: I0930 15:40:39.593355 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f6f049b3-7415-4acc-829a-659677059137/setup-container/0.log" Sep 30 15:40:40 crc kubenswrapper[4799]: I0930 15:40:40.443930 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f6f049b3-7415-4acc-829a-659677059137/setup-container/0.log" Sep 30 15:40:40 crc kubenswrapper[4799]: I0930 15:40:40.651397 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2/setup-container/0.log" Sep 30 15:40:40 crc kubenswrapper[4799]: I0930 15:40:40.690728 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f6f049b3-7415-4acc-829a-659677059137/rabbitmq/0.log" Sep 30 15:40:41 crc kubenswrapper[4799]: I0930 15:40:41.104402 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2/setup-container/0.log" Sep 30 15:40:41 crc kubenswrapper[4799]: I0930 15:40:41.163159 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2/rabbitmq/0.log" Sep 30 15:40:41 crc kubenswrapper[4799]: I0930 15:40:41.222125 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n_aadd9e81-92aa-46fe-bae7-8a0ccddd873f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:41 crc kubenswrapper[4799]: I0930 15:40:41.362971 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-nwf55_286e79d0-072f-422d-82cf-8932a5e9839f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:41 crc kubenswrapper[4799]: I0930 15:40:41.473986 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7_2fdc7b08-b5e8-45cb-97d9-d44658909edf/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:41 crc kubenswrapper[4799]: I0930 15:40:41.708393 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-g782p_34ff6ecd-36de-4642-953a-c5d60e8ab31e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:41 crc kubenswrapper[4799]: I0930 15:40:41.798468 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8swsg_5f994a1a-5c13-4a9e-9d44-ea7a8389e943/ssh-known-hosts-edpm-deployment/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.067290 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5cfc9f87d7-nj7t8_f32c852d-173a-4417-a268-38b7a480937e/proxy-server/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.076805 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5cfc9f87d7-nj7t8_f32c852d-173a-4417-a268-38b7a480937e/proxy-httpd/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.229121 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-mbljj_ad219d29-0576-4b77-b5b9-fc2c980b7093/swift-ring-rebalance/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.304919 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/account-auditor/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.330018 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/account-reaper/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.454353 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/account-replicator/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.567508 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/account-server/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.574686 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/container-auditor/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.663872 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/container-replicator/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.788877 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/container-server/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.837512 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/container-updater/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.940240 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-auditor/0.log" Sep 30 15:40:42 crc kubenswrapper[4799]: I0930 15:40:42.940618 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-expirer/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.027785 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-server/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.118354 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-replicator/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.192859 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/rsync/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.200411 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-updater/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.337119 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/swift-recon-cron/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.483466 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-f72vr_a86a2d67-7c0c-44b5-a435-15fccfd80d5e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.614265 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ab210e62-906f-46ad-b80a-fbf00a875124/tempest-tests-tempest-tests-runner/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.688286 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b3f82270-84d2-409f-b243-cde51bfb251f/test-operator-logs-container/0.log" Sep 30 15:40:43 crc kubenswrapper[4799]: I0930 15:40:43.876659 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9p67d_06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:40:59 crc kubenswrapper[4799]: I0930 15:40:59.648913 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:40:59 crc kubenswrapper[4799]: I0930 15:40:59.652851 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:40:59 crc kubenswrapper[4799]: I0930 15:40:59.653054 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:40:59 crc kubenswrapper[4799]: I0930 15:40:59.654149 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4b97db586cc9e0bbe8e063b258329d9e848a6cab8be98493164f05db56b329d"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:40:59 crc kubenswrapper[4799]: I0930 15:40:59.654305 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://e4b97db586cc9e0bbe8e063b258329d9e848a6cab8be98493164f05db56b329d" gracePeriod=600 Sep 30 15:40:59 crc kubenswrapper[4799]: I0930 15:40:59.920557 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="e4b97db586cc9e0bbe8e063b258329d9e848a6cab8be98493164f05db56b329d" exitCode=0 Sep 30 15:40:59 crc kubenswrapper[4799]: I0930 15:40:59.920608 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"e4b97db586cc9e0bbe8e063b258329d9e848a6cab8be98493164f05db56b329d"} Sep 30 15:40:59 crc kubenswrapper[4799]: I0930 15:40:59.920704 4799 scope.go:117] "RemoveContainer" containerID="d4ba9587dbda199c0c508f7f4ab1d8267f436a9b8fde9cb569083fb066882a12" Sep 30 15:41:00 crc kubenswrapper[4799]: I0930 15:41:00.932334 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109"} Sep 30 15:41:34 crc kubenswrapper[4799]: E0930 15:41:34.368334 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88ef93f5_1f51_4863_bf88_57e75d415b68.slice/crio-conmon-997a7fd3789451e743f96887058dd1ab8ea90dcb13c08271a8f88c8f0ef9e903.scope\": RecentStats: unable to find data in memory cache]" Sep 30 15:41:35 crc kubenswrapper[4799]: I0930 15:41:35.267320 4799 generic.go:334] "Generic (PLEG): container finished" podID="88ef93f5-1f51-4863-bf88-57e75d415b68" containerID="997a7fd3789451e743f96887058dd1ab8ea90dcb13c08271a8f88c8f0ef9e903" exitCode=0 Sep 30 15:41:35 crc kubenswrapper[4799]: I0930 15:41:35.267521 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" event={"ID":"88ef93f5-1f51-4863-bf88-57e75d415b68","Type":"ContainerDied","Data":"997a7fd3789451e743f96887058dd1ab8ea90dcb13c08271a8f88c8f0ef9e903"} Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.419411 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.518854 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9psmp\" (UniqueName: \"kubernetes.io/projected/88ef93f5-1f51-4863-bf88-57e75d415b68-kube-api-access-9psmp\") pod \"88ef93f5-1f51-4863-bf88-57e75d415b68\" (UID: \"88ef93f5-1f51-4863-bf88-57e75d415b68\") " Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.520823 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ef93f5-1f51-4863-bf88-57e75d415b68-host\") pod \"88ef93f5-1f51-4863-bf88-57e75d415b68\" (UID: \"88ef93f5-1f51-4863-bf88-57e75d415b68\") " Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.520927 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88ef93f5-1f51-4863-bf88-57e75d415b68-host" (OuterVolumeSpecName: "host") pod "88ef93f5-1f51-4863-bf88-57e75d415b68" (UID: "88ef93f5-1f51-4863-bf88-57e75d415b68"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.521537 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/88ef93f5-1f51-4863-bf88-57e75d415b68-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.530146 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ef93f5-1f51-4863-bf88-57e75d415b68-kube-api-access-9psmp" (OuterVolumeSpecName: "kube-api-access-9psmp") pod "88ef93f5-1f51-4863-bf88-57e75d415b68" (UID: "88ef93f5-1f51-4863-bf88-57e75d415b68"). InnerVolumeSpecName "kube-api-access-9psmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.599951 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-5tkj9"] Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.612007 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-5tkj9"] Sep 30 15:41:36 crc kubenswrapper[4799]: I0930 15:41:36.624474 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9psmp\" (UniqueName: \"kubernetes.io/projected/88ef93f5-1f51-4863-bf88-57e75d415b68-kube-api-access-9psmp\") on node \"crc\" DevicePath \"\"" Sep 30 15:41:37 crc kubenswrapper[4799]: I0930 15:41:37.293282 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff96424c2a13565862df532e8c937f52a329f0fa5f53ccdc7342ccd7de04ac1d" Sep 30 15:41:37 crc kubenswrapper[4799]: I0930 15:41:37.293325 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-5tkj9" Sep 30 15:41:37 crc kubenswrapper[4799]: I0930 15:41:37.907000 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-99lmt"] Sep 30 15:41:37 crc kubenswrapper[4799]: E0930 15:41:37.907746 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ef93f5-1f51-4863-bf88-57e75d415b68" containerName="container-00" Sep 30 15:41:37 crc kubenswrapper[4799]: I0930 15:41:37.907764 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ef93f5-1f51-4863-bf88-57e75d415b68" containerName="container-00" Sep 30 15:41:37 crc kubenswrapper[4799]: I0930 15:41:37.908015 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ef93f5-1f51-4863-bf88-57e75d415b68" containerName="container-00" Sep 30 15:41:37 crc kubenswrapper[4799]: I0930 15:41:37.908795 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.052958 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr9fq\" (UniqueName: \"kubernetes.io/projected/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-kube-api-access-qr9fq\") pod \"crc-debug-99lmt\" (UID: \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\") " pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.053084 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-host\") pod \"crc-debug-99lmt\" (UID: \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\") " pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.155580 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-host\") pod \"crc-debug-99lmt\" (UID: \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\") " pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.155770 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr9fq\" (UniqueName: \"kubernetes.io/projected/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-kube-api-access-qr9fq\") pod \"crc-debug-99lmt\" (UID: \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\") " pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.155781 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-host\") pod \"crc-debug-99lmt\" (UID: \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\") " pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.175714 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr9fq\" (UniqueName: \"kubernetes.io/projected/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-kube-api-access-qr9fq\") pod \"crc-debug-99lmt\" (UID: \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\") " pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.225413 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.303505 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" event={"ID":"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84","Type":"ContainerStarted","Data":"8466ff3ddffb525bd380b25b469ff50ed1a70d318eaec6c2bfcf5b7282f03192"} Sep 30 15:41:38 crc kubenswrapper[4799]: I0930 15:41:38.514966 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ef93f5-1f51-4863-bf88-57e75d415b68" path="/var/lib/kubelet/pods/88ef93f5-1f51-4863-bf88-57e75d415b68/volumes" Sep 30 15:41:39 crc kubenswrapper[4799]: I0930 15:41:39.317269 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" event={"ID":"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84","Type":"ContainerStarted","Data":"976f783642389069ea60d2c6e2a7ba340ee537cb9916c25d5ae6a3ca3692abc0"} Sep 30 15:41:39 crc kubenswrapper[4799]: I0930 15:41:39.341191 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" podStartSLOduration=2.3411654029999998 podStartE2EDuration="2.341165403s" podCreationTimestamp="2025-09-30 15:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:41:39.33249854 +0000 UTC m=+4921.416098987" watchObservedRunningTime="2025-09-30 15:41:39.341165403 +0000 UTC m=+4921.424765830" Sep 30 15:41:40 crc kubenswrapper[4799]: I0930 15:41:40.329809 4799 generic.go:334] "Generic (PLEG): container finished" podID="c9d0a9ac-476e-4890-96dd-6a94ea7d0b84" containerID="976f783642389069ea60d2c6e2a7ba340ee537cb9916c25d5ae6a3ca3692abc0" exitCode=0 Sep 30 15:41:40 crc kubenswrapper[4799]: I0930 15:41:40.329859 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" event={"ID":"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84","Type":"ContainerDied","Data":"976f783642389069ea60d2c6e2a7ba340ee537cb9916c25d5ae6a3ca3692abc0"} Sep 30 15:41:41 crc kubenswrapper[4799]: I0930 15:41:41.456623 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:41 crc kubenswrapper[4799]: I0930 15:41:41.532852 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr9fq\" (UniqueName: \"kubernetes.io/projected/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-kube-api-access-qr9fq\") pod \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\" (UID: \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\") " Sep 30 15:41:41 crc kubenswrapper[4799]: I0930 15:41:41.533100 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-host\") pod \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\" (UID: \"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84\") " Sep 30 15:41:41 crc kubenswrapper[4799]: I0930 15:41:41.533772 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-host" (OuterVolumeSpecName: "host") pod "c9d0a9ac-476e-4890-96dd-6a94ea7d0b84" (UID: "c9d0a9ac-476e-4890-96dd-6a94ea7d0b84"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:41:41 crc kubenswrapper[4799]: I0930 15:41:41.558927 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-kube-api-access-qr9fq" (OuterVolumeSpecName: "kube-api-access-qr9fq") pod "c9d0a9ac-476e-4890-96dd-6a94ea7d0b84" (UID: "c9d0a9ac-476e-4890-96dd-6a94ea7d0b84"). InnerVolumeSpecName "kube-api-access-qr9fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:41:41 crc kubenswrapper[4799]: I0930 15:41:41.637077 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr9fq\" (UniqueName: \"kubernetes.io/projected/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-kube-api-access-qr9fq\") on node \"crc\" DevicePath \"\"" Sep 30 15:41:41 crc kubenswrapper[4799]: I0930 15:41:41.638569 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:41:42 crc kubenswrapper[4799]: I0930 15:41:42.357888 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" event={"ID":"c9d0a9ac-476e-4890-96dd-6a94ea7d0b84","Type":"ContainerDied","Data":"8466ff3ddffb525bd380b25b469ff50ed1a70d318eaec6c2bfcf5b7282f03192"} Sep 30 15:41:42 crc kubenswrapper[4799]: I0930 15:41:42.357947 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8466ff3ddffb525bd380b25b469ff50ed1a70d318eaec6c2bfcf5b7282f03192" Sep 30 15:41:42 crc kubenswrapper[4799]: I0930 15:41:42.357962 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-99lmt" Sep 30 15:41:48 crc kubenswrapper[4799]: I0930 15:41:48.970981 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-99lmt"] Sep 30 15:41:48 crc kubenswrapper[4799]: I0930 15:41:48.981952 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-99lmt"] Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.514589 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9d0a9ac-476e-4890-96dd-6a94ea7d0b84" path="/var/lib/kubelet/pods/c9d0a9ac-476e-4890-96dd-6a94ea7d0b84/volumes" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.574696 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-8m86d"] Sep 30 15:41:50 crc kubenswrapper[4799]: E0930 15:41:50.575175 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9d0a9ac-476e-4890-96dd-6a94ea7d0b84" containerName="container-00" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.575196 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9d0a9ac-476e-4890-96dd-6a94ea7d0b84" containerName="container-00" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.575440 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9d0a9ac-476e-4890-96dd-6a94ea7d0b84" containerName="container-00" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.576176 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.719981 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/084d1c88-ef33-40ff-883c-fb3b063e82fe-host\") pod \"crc-debug-8m86d\" (UID: \"084d1c88-ef33-40ff-883c-fb3b063e82fe\") " pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.720601 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9jwh\" (UniqueName: \"kubernetes.io/projected/084d1c88-ef33-40ff-883c-fb3b063e82fe-kube-api-access-f9jwh\") pod \"crc-debug-8m86d\" (UID: \"084d1c88-ef33-40ff-883c-fb3b063e82fe\") " pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.822713 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/084d1c88-ef33-40ff-883c-fb3b063e82fe-host\") pod \"crc-debug-8m86d\" (UID: \"084d1c88-ef33-40ff-883c-fb3b063e82fe\") " pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.822807 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9jwh\" (UniqueName: \"kubernetes.io/projected/084d1c88-ef33-40ff-883c-fb3b063e82fe-kube-api-access-f9jwh\") pod \"crc-debug-8m86d\" (UID: \"084d1c88-ef33-40ff-883c-fb3b063e82fe\") " pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.822927 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/084d1c88-ef33-40ff-883c-fb3b063e82fe-host\") pod \"crc-debug-8m86d\" (UID: \"084d1c88-ef33-40ff-883c-fb3b063e82fe\") " pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.851644 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9jwh\" (UniqueName: \"kubernetes.io/projected/084d1c88-ef33-40ff-883c-fb3b063e82fe-kube-api-access-f9jwh\") pod \"crc-debug-8m86d\" (UID: \"084d1c88-ef33-40ff-883c-fb3b063e82fe\") " pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:50 crc kubenswrapper[4799]: I0930 15:41:50.902785 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:50 crc kubenswrapper[4799]: W0930 15:41:50.946870 4799 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod084d1c88_ef33_40ff_883c_fb3b063e82fe.slice/crio-6cb53b136c12052b2f97f5fd39201fe74205ba8c88c7ea7871150822c7260308 WatchSource:0}: Error finding container 6cb53b136c12052b2f97f5fd39201fe74205ba8c88c7ea7871150822c7260308: Status 404 returned error can't find the container with id 6cb53b136c12052b2f97f5fd39201fe74205ba8c88c7ea7871150822c7260308 Sep 30 15:41:51 crc kubenswrapper[4799]: I0930 15:41:51.457337 4799 generic.go:334] "Generic (PLEG): container finished" podID="084d1c88-ef33-40ff-883c-fb3b063e82fe" containerID="2e24090e494ac7b902f4f19bb078d325bd3698e62b72968173d186317e2378fd" exitCode=0 Sep 30 15:41:51 crc kubenswrapper[4799]: I0930 15:41:51.457407 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-8m86d" event={"ID":"084d1c88-ef33-40ff-883c-fb3b063e82fe","Type":"ContainerDied","Data":"2e24090e494ac7b902f4f19bb078d325bd3698e62b72968173d186317e2378fd"} Sep 30 15:41:51 crc kubenswrapper[4799]: I0930 15:41:51.457830 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/crc-debug-8m86d" event={"ID":"084d1c88-ef33-40ff-883c-fb3b063e82fe","Type":"ContainerStarted","Data":"6cb53b136c12052b2f97f5fd39201fe74205ba8c88c7ea7871150822c7260308"} Sep 30 15:41:51 crc kubenswrapper[4799]: I0930 15:41:51.531355 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-8m86d"] Sep 30 15:41:51 crc kubenswrapper[4799]: I0930 15:41:51.542739 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d2c2l/crc-debug-8m86d"] Sep 30 15:41:52 crc kubenswrapper[4799]: I0930 15:41:52.586921 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:52 crc kubenswrapper[4799]: I0930 15:41:52.672215 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9jwh\" (UniqueName: \"kubernetes.io/projected/084d1c88-ef33-40ff-883c-fb3b063e82fe-kube-api-access-f9jwh\") pod \"084d1c88-ef33-40ff-883c-fb3b063e82fe\" (UID: \"084d1c88-ef33-40ff-883c-fb3b063e82fe\") " Sep 30 15:41:52 crc kubenswrapper[4799]: I0930 15:41:52.672625 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/084d1c88-ef33-40ff-883c-fb3b063e82fe-host\") pod \"084d1c88-ef33-40ff-883c-fb3b063e82fe\" (UID: \"084d1c88-ef33-40ff-883c-fb3b063e82fe\") " Sep 30 15:41:52 crc kubenswrapper[4799]: I0930 15:41:52.673061 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/084d1c88-ef33-40ff-883c-fb3b063e82fe-host" (OuterVolumeSpecName: "host") pod "084d1c88-ef33-40ff-883c-fb3b063e82fe" (UID: "084d1c88-ef33-40ff-883c-fb3b063e82fe"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:41:52 crc kubenswrapper[4799]: I0930 15:41:52.690965 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/084d1c88-ef33-40ff-883c-fb3b063e82fe-kube-api-access-f9jwh" (OuterVolumeSpecName: "kube-api-access-f9jwh") pod "084d1c88-ef33-40ff-883c-fb3b063e82fe" (UID: "084d1c88-ef33-40ff-883c-fb3b063e82fe"). InnerVolumeSpecName "kube-api-access-f9jwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:41:52 crc kubenswrapper[4799]: I0930 15:41:52.774633 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/084d1c88-ef33-40ff-883c-fb3b063e82fe-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:41:52 crc kubenswrapper[4799]: I0930 15:41:52.774694 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9jwh\" (UniqueName: \"kubernetes.io/projected/084d1c88-ef33-40ff-883c-fb3b063e82fe-kube-api-access-f9jwh\") on node \"crc\" DevicePath \"\"" Sep 30 15:41:53 crc kubenswrapper[4799]: I0930 15:41:53.459073 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/util/0.log" Sep 30 15:41:53 crc kubenswrapper[4799]: I0930 15:41:53.477858 4799 scope.go:117] "RemoveContainer" containerID="2e24090e494ac7b902f4f19bb078d325bd3698e62b72968173d186317e2378fd" Sep 30 15:41:53 crc kubenswrapper[4799]: I0930 15:41:53.477936 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/crc-debug-8m86d" Sep 30 15:41:53 crc kubenswrapper[4799]: I0930 15:41:53.820899 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/pull/0.log" Sep 30 15:41:53 crc kubenswrapper[4799]: I0930 15:41:53.821035 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/pull/0.log" Sep 30 15:41:53 crc kubenswrapper[4799]: I0930 15:41:53.852314 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/util/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.015496 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/pull/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.071479 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/util/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.110023 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/extract/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.265774 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-j7tgf_b8a734fd-7d92-4b2d-9bb4-91eee56047c5/kube-rbac-proxy/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.373803 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-j7tgf_b8a734fd-7d92-4b2d-9bb4-91eee56047c5/manager/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.410558 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kqfcp_e0002170-fb28-4c87-a970-350f92c891bd/kube-rbac-proxy/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.520615 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="084d1c88-ef33-40ff-883c-fb3b063e82fe" path="/var/lib/kubelet/pods/084d1c88-ef33-40ff-883c-fb3b063e82fe/volumes" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.645046 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kqfcp_e0002170-fb28-4c87-a970-350f92c891bd/manager/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.701036 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-mplqn_8bc24269-6185-41a5-861b-19c88576c223/kube-rbac-proxy/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.734597 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-mplqn_8bc24269-6185-41a5-861b-19c88576c223/manager/0.log" Sep 30 15:41:54 crc kubenswrapper[4799]: I0930 15:41:54.931537 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-gmxsp_19539cfa-6f83-4636-8193-d2373038f353/kube-rbac-proxy/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.090995 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-gmxsp_19539cfa-6f83-4636-8193-d2373038f353/manager/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.176520 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-8vxq2_08704fb7-1721-4f10-97a1-f2359d01cdaf/kube-rbac-proxy/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.220741 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-8vxq2_08704fb7-1721-4f10-97a1-f2359d01cdaf/manager/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.287942 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-k955x_fdb1977f-1341-46ae-a4db-e78e20b6d5bd/kube-rbac-proxy/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.499390 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-k955x_fdb1977f-1341-46ae-a4db-e78e20b6d5bd/manager/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.606391 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-rzjpm_cdca0d27-1d59-4858-85a4-681c1d6abf90/kube-rbac-proxy/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.740386 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-p75vq_2e07068f-3f60-4b35-9465-1adb7cda6525/kube-rbac-proxy/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.841396 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-p75vq_2e07068f-3f60-4b35-9465-1adb7cda6525/manager/0.log" Sep 30 15:41:55 crc kubenswrapper[4799]: I0930 15:41:55.901634 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-rzjpm_cdca0d27-1d59-4858-85a4-681c1d6abf90/manager/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.128328 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-vq46v_943ee0ba-2610-4d88-b211-4ddf83cb5b03/manager/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.149363 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-vq46v_943ee0ba-2610-4d88-b211-4ddf83cb5b03/kube-rbac-proxy/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.355901 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-2jnc8_40600429-6eda-47b1-bf61-0e22189f7549/kube-rbac-proxy/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.404522 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-2jnc8_40600429-6eda-47b1-bf61-0e22189f7549/manager/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.481440 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-pxd4j_c8d8a2c1-6828-49d3-bd9a-baa3fe58790e/kube-rbac-proxy/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.646135 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-pxd4j_c8d8a2c1-6828-49d3-bd9a-baa3fe58790e/manager/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.718368 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-9mvns_f3e1a985-b655-4dc8-8e5f-f49a6459f512/kube-rbac-proxy/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.780492 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-9mvns_f3e1a985-b655-4dc8-8e5f-f49a6459f512/manager/0.log" Sep 30 15:41:56 crc kubenswrapper[4799]: I0930 15:41:56.905900 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-vwjtl_4ecfaaa8-a63c-4f41-a08b-5d497a71e16d/kube-rbac-proxy/0.log" Sep 30 15:41:57 crc kubenswrapper[4799]: I0930 15:41:57.094224 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-vwjtl_4ecfaaa8-a63c-4f41-a08b-5d497a71e16d/manager/0.log" Sep 30 15:41:57 crc kubenswrapper[4799]: I0930 15:41:57.146424 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-rpsrk_34ffd2cc-7377-4160-b03d-cae022be0262/kube-rbac-proxy/0.log" Sep 30 15:41:57 crc kubenswrapper[4799]: I0930 15:41:57.286115 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-rpsrk_34ffd2cc-7377-4160-b03d-cae022be0262/manager/0.log" Sep 30 15:41:57 crc kubenswrapper[4799]: I0930 15:41:57.372011 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-q7tzd_68487038-29af-45ea-8619-acd75f3d9a7d/kube-rbac-proxy/0.log" Sep 30 15:41:57 crc kubenswrapper[4799]: I0930 15:41:57.482129 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-q7tzd_68487038-29af-45ea-8619-acd75f3d9a7d/manager/0.log" Sep 30 15:41:57 crc kubenswrapper[4799]: I0930 15:41:57.535351 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-698bb85c6f-m82xh_28ece3c4-17d4-4b6f-8156-f90cfb794581/kube-rbac-proxy/0.log" Sep 30 15:41:57 crc kubenswrapper[4799]: I0930 15:41:57.779185 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-655f477c67-z5l5q_2cfc3242-481e-4575-9d3a-bbd740b097f6/kube-rbac-proxy/0.log" Sep 30 15:41:58 crc kubenswrapper[4799]: I0930 15:41:58.050348 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-w6tbf_f2bdc984-2b22-4a12-99b0-ef6235691241/registry-server/0.log" Sep 30 15:41:58 crc kubenswrapper[4799]: I0930 15:41:58.112597 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-655f477c67-z5l5q_2cfc3242-481e-4575-9d3a-bbd740b097f6/operator/0.log" Sep 30 15:41:58 crc kubenswrapper[4799]: I0930 15:41:58.223759 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-qrwt4_ee68b34b-53e7-452e-a113-ed35ba5c0b85/kube-rbac-proxy/0.log" Sep 30 15:41:58 crc kubenswrapper[4799]: I0930 15:41:58.443825 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-sz2rp_b9a9cd33-4400-4568-822f-effb32929bb3/kube-rbac-proxy/0.log" Sep 30 15:41:58 crc kubenswrapper[4799]: I0930 15:41:58.481214 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-qrwt4_ee68b34b-53e7-452e-a113-ed35ba5c0b85/manager/0.log" Sep 30 15:41:58 crc kubenswrapper[4799]: I0930 15:41:58.497840 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-sz2rp_b9a9cd33-4400-4568-822f-effb32929bb3/manager/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.259375 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-7tlqx_78d18534-fd0c-454f-80f3-b65ddd5a24c9/operator/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.321977 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-68qzx_208f263c-4523-45e3-bf7e-19c8407bc599/kube-rbac-proxy/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.348181 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-698bb85c6f-m82xh_28ece3c4-17d4-4b6f-8156-f90cfb794581/manager/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.539153 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-68qzx_208f263c-4523-45e3-bf7e-19c8407bc599/manager/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.570436 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-9mwxx_3a3cf54d-3dce-46a0-873c-412159358f6a/kube-rbac-proxy/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.650928 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-9mwxx_3a3cf54d-3dce-46a0-873c-412159358f6a/manager/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.777190 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-nxbv9_bb413bd9-a2c6-43c9-95d4-c59bbc08e930/kube-rbac-proxy/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.792615 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-nxbv9_bb413bd9-a2c6-43c9-95d4-c59bbc08e930/manager/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.877306 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-c7mtd_26c4c0df-9057-493d-9a60-b39c664991a4/kube-rbac-proxy/0.log" Sep 30 15:41:59 crc kubenswrapper[4799]: I0930 15:41:59.924751 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-c7mtd_26c4c0df-9057-493d-9a60-b39c664991a4/manager/0.log" Sep 30 15:42:23 crc kubenswrapper[4799]: I0930 15:42:23.179291 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j5jqc_7c44311c-64e8-4587-95c6-2bdda1c93ad1/control-plane-machine-set-operator/0.log" Sep 30 15:42:23 crc kubenswrapper[4799]: I0930 15:42:23.392576 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jgd4t_018cb267-3656-4202-ba7e-323811533a19/machine-api-operator/0.log" Sep 30 15:42:23 crc kubenswrapper[4799]: I0930 15:42:23.397951 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jgd4t_018cb267-3656-4202-ba7e-323811533a19/kube-rbac-proxy/0.log" Sep 30 15:42:37 crc kubenswrapper[4799]: I0930 15:42:37.453418 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-mn5dh_e3b68d27-f024-4124-99d3-d16f2171c911/cert-manager-controller/0.log" Sep 30 15:42:37 crc kubenswrapper[4799]: I0930 15:42:37.783903 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-vdjtl_9146b875-e4c0-451d-8991-858332cc39d7/cert-manager-cainjector/0.log" Sep 30 15:42:37 crc kubenswrapper[4799]: I0930 15:42:37.837499 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-94zcs_8bb043b3-1966-4b5e-bc13-476e14eb927e/cert-manager-webhook/0.log" Sep 30 15:42:50 crc kubenswrapper[4799]: I0930 15:42:50.925990 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-pbz2f_3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69/nmstate-console-plugin/0.log" Sep 30 15:42:51 crc kubenswrapper[4799]: I0930 15:42:51.608067 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-bk8lx_dc69b189-0b92-42fb-9584-42fd5aed183f/kube-rbac-proxy/0.log" Sep 30 15:42:51 crc kubenswrapper[4799]: I0930 15:42:51.614363 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-672fh_60de70ea-f8ce-4f5a-ab69-45fe7cd77e20/nmstate-handler/0.log" Sep 30 15:42:51 crc kubenswrapper[4799]: I0930 15:42:51.741714 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-bk8lx_dc69b189-0b92-42fb-9584-42fd5aed183f/nmstate-metrics/0.log" Sep 30 15:42:51 crc kubenswrapper[4799]: I0930 15:42:51.824824 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-hh9gx_311ed4a0-c6f5-459d-9da6-86dbe84dd12f/nmstate-operator/0.log" Sep 30 15:42:51 crc kubenswrapper[4799]: I0930 15:42:51.940148 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-jdngw_23e26176-f572-43f8-baba-8bbe27fa0bca/nmstate-webhook/0.log" Sep 30 15:43:06 crc kubenswrapper[4799]: I0930 15:43:06.743266 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-dcxwg_96370b88-b581-4342-a570-059733689d3e/kube-rbac-proxy/0.log" Sep 30 15:43:06 crc kubenswrapper[4799]: I0930 15:43:06.972458 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-dcxwg_96370b88-b581-4342-a570-059733689d3e/controller/0.log" Sep 30 15:43:06 crc kubenswrapper[4799]: I0930 15:43:06.978755 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-frr-files/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.193704 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-frr-files/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.284032 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-reloader/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.314966 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-reloader/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.315118 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-metrics/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.553750 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-metrics/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.561460 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-metrics/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.586250 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-frr-files/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.618303 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-reloader/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.794188 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-frr-files/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.850767 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-reloader/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.894088 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-metrics/0.log" Sep 30 15:43:07 crc kubenswrapper[4799]: I0930 15:43:07.917633 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/controller/0.log" Sep 30 15:43:08 crc kubenswrapper[4799]: I0930 15:43:08.070596 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/kube-rbac-proxy/0.log" Sep 30 15:43:08 crc kubenswrapper[4799]: I0930 15:43:08.246601 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/kube-rbac-proxy-frr/0.log" Sep 30 15:43:08 crc kubenswrapper[4799]: I0930 15:43:08.249775 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/frr-metrics/0.log" Sep 30 15:43:08 crc kubenswrapper[4799]: I0930 15:43:08.429174 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/reloader/0.log" Sep 30 15:43:08 crc kubenswrapper[4799]: I0930 15:43:08.571569 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-z8j6r_17444fcf-15e8-4038-9500-508d084a38cd/frr-k8s-webhook-server/0.log" Sep 30 15:43:08 crc kubenswrapper[4799]: I0930 15:43:08.883994 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-668b984c56-cwmvh_72f552b4-b1cb-4659-8629-d8ff13a85cc5/manager/0.log" Sep 30 15:43:09 crc kubenswrapper[4799]: I0930 15:43:09.097061 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-698f89f9d8-jdx5j_624ab48f-5623-4bf1-af7f-4cad0be3e654/webhook-server/0.log" Sep 30 15:43:09 crc kubenswrapper[4799]: I0930 15:43:09.271925 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mtwhr_056d4eef-c8d4-4f76-9006-cb6454ec953c/kube-rbac-proxy/0.log" Sep 30 15:43:09 crc kubenswrapper[4799]: I0930 15:43:09.570908 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/frr/0.log" Sep 30 15:43:09 crc kubenswrapper[4799]: I0930 15:43:09.799057 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mtwhr_056d4eef-c8d4-4f76-9006-cb6454ec953c/speaker/0.log" Sep 30 15:43:23 crc kubenswrapper[4799]: I0930 15:43:23.823818 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/util/0.log" Sep 30 15:43:23 crc kubenswrapper[4799]: I0930 15:43:23.948178 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/util/0.log" Sep 30 15:43:24 crc kubenswrapper[4799]: I0930 15:43:24.020275 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/pull/0.log" Sep 30 15:43:24 crc kubenswrapper[4799]: I0930 15:43:24.050741 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/pull/0.log" Sep 30 15:43:24 crc kubenswrapper[4799]: I0930 15:43:24.213273 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/util/0.log" Sep 30 15:43:24 crc kubenswrapper[4799]: I0930 15:43:24.222098 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/extract/0.log" Sep 30 15:43:24 crc kubenswrapper[4799]: I0930 15:43:24.264143 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/pull/0.log" Sep 30 15:43:24 crc kubenswrapper[4799]: I0930 15:43:24.468969 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-utilities/0.log" Sep 30 15:43:25 crc kubenswrapper[4799]: I0930 15:43:25.154909 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-content/0.log" Sep 30 15:43:25 crc kubenswrapper[4799]: I0930 15:43:25.178233 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-content/0.log" Sep 30 15:43:25 crc kubenswrapper[4799]: I0930 15:43:25.198242 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-utilities/0.log" Sep 30 15:43:25 crc kubenswrapper[4799]: I0930 15:43:25.514723 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-content/0.log" Sep 30 15:43:25 crc kubenswrapper[4799]: I0930 15:43:25.553081 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-utilities/0.log" Sep 30 15:43:26 crc kubenswrapper[4799]: I0930 15:43:26.171864 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-utilities/0.log" Sep 30 15:43:26 crc kubenswrapper[4799]: I0930 15:43:26.581205 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/registry-server/0.log" Sep 30 15:43:26 crc kubenswrapper[4799]: I0930 15:43:26.911709 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-content/0.log" Sep 30 15:43:26 crc kubenswrapper[4799]: I0930 15:43:26.959729 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-utilities/0.log" Sep 30 15:43:26 crc kubenswrapper[4799]: I0930 15:43:26.977447 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-content/0.log" Sep 30 15:43:27 crc kubenswrapper[4799]: I0930 15:43:27.159564 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-utilities/0.log" Sep 30 15:43:27 crc kubenswrapper[4799]: I0930 15:43:27.175535 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-content/0.log" Sep 30 15:43:27 crc kubenswrapper[4799]: I0930 15:43:27.527009 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/util/0.log" Sep 30 15:43:27 crc kubenswrapper[4799]: I0930 15:43:27.650491 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/registry-server/0.log" Sep 30 15:43:27 crc kubenswrapper[4799]: I0930 15:43:27.894380 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/pull/0.log" Sep 30 15:43:27 crc kubenswrapper[4799]: I0930 15:43:27.924047 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/util/0.log" Sep 30 15:43:27 crc kubenswrapper[4799]: I0930 15:43:27.928673 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/pull/0.log" Sep 30 15:43:28 crc kubenswrapper[4799]: I0930 15:43:28.257058 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/util/0.log" Sep 30 15:43:28 crc kubenswrapper[4799]: I0930 15:43:28.274809 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/pull/0.log" Sep 30 15:43:28 crc kubenswrapper[4799]: I0930 15:43:28.275184 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/extract/0.log" Sep 30 15:43:28 crc kubenswrapper[4799]: I0930 15:43:28.349525 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z2qlr_7f71a790-5180-4e2c-be45-837154fa1d62/marketplace-operator/0.log" Sep 30 15:43:28 crc kubenswrapper[4799]: I0930 15:43:28.518954 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-utilities/0.log" Sep 30 15:43:28 crc kubenswrapper[4799]: I0930 15:43:28.777390 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-utilities/0.log" Sep 30 15:43:28 crc kubenswrapper[4799]: I0930 15:43:28.785858 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-content/0.log" Sep 30 15:43:28 crc kubenswrapper[4799]: I0930 15:43:28.816437 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-content/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.051139 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-utilities/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.130466 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-utilities/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.218470 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-content/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.244198 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/registry-server/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.448014 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-content/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.504534 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-utilities/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.559505 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-content/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.648542 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.648670 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.780945 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-content/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.790321 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-utilities/0.log" Sep 30 15:43:29 crc kubenswrapper[4799]: I0930 15:43:29.901546 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/registry-server/0.log" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.405045 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wf5h6"] Sep 30 15:43:52 crc kubenswrapper[4799]: E0930 15:43:52.407900 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="084d1c88-ef33-40ff-883c-fb3b063e82fe" containerName="container-00" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.408040 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="084d1c88-ef33-40ff-883c-fb3b063e82fe" containerName="container-00" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.408404 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="084d1c88-ef33-40ff-883c-fb3b063e82fe" containerName="container-00" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.410412 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.435502 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wf5h6"] Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.580916 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvffc\" (UniqueName: \"kubernetes.io/projected/57a77b0b-5d24-4e9d-9a34-aff8d8757873-kube-api-access-zvffc\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.581360 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-catalog-content\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.581421 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-utilities\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.683283 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvffc\" (UniqueName: \"kubernetes.io/projected/57a77b0b-5d24-4e9d-9a34-aff8d8757873-kube-api-access-zvffc\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.683368 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-catalog-content\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.683417 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-utilities\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.683974 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-catalog-content\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.684032 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-utilities\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.720705 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvffc\" (UniqueName: \"kubernetes.io/projected/57a77b0b-5d24-4e9d-9a34-aff8d8757873-kube-api-access-zvffc\") pod \"certified-operators-wf5h6\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:52 crc kubenswrapper[4799]: I0930 15:43:52.731695 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:43:53 crc kubenswrapper[4799]: I0930 15:43:53.854855 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wf5h6"] Sep 30 15:43:54 crc kubenswrapper[4799]: I0930 15:43:54.674459 4799 generic.go:334] "Generic (PLEG): container finished" podID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerID="c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c" exitCode=0 Sep 30 15:43:54 crc kubenswrapper[4799]: I0930 15:43:54.674514 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wf5h6" event={"ID":"57a77b0b-5d24-4e9d-9a34-aff8d8757873","Type":"ContainerDied","Data":"c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c"} Sep 30 15:43:54 crc kubenswrapper[4799]: I0930 15:43:54.674770 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wf5h6" event={"ID":"57a77b0b-5d24-4e9d-9a34-aff8d8757873","Type":"ContainerStarted","Data":"00cc15b5c7d726db89c4885816ab04afd5447f9a2be37a3669a7df39c3a3d8d7"} Sep 30 15:43:54 crc kubenswrapper[4799]: I0930 15:43:54.687830 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 15:43:56 crc kubenswrapper[4799]: I0930 15:43:56.700855 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wf5h6" event={"ID":"57a77b0b-5d24-4e9d-9a34-aff8d8757873","Type":"ContainerStarted","Data":"345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e"} Sep 30 15:43:57 crc kubenswrapper[4799]: I0930 15:43:57.720859 4799 generic.go:334] "Generic (PLEG): container finished" podID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerID="345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e" exitCode=0 Sep 30 15:43:57 crc kubenswrapper[4799]: I0930 15:43:57.721226 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wf5h6" event={"ID":"57a77b0b-5d24-4e9d-9a34-aff8d8757873","Type":"ContainerDied","Data":"345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e"} Sep 30 15:43:59 crc kubenswrapper[4799]: I0930 15:43:59.652233 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:43:59 crc kubenswrapper[4799]: I0930 15:43:59.654011 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:43:59 crc kubenswrapper[4799]: I0930 15:43:59.743132 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wf5h6" event={"ID":"57a77b0b-5d24-4e9d-9a34-aff8d8757873","Type":"ContainerStarted","Data":"751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866"} Sep 30 15:43:59 crc kubenswrapper[4799]: I0930 15:43:59.763014 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wf5h6" podStartSLOduration=4.166408184 podStartE2EDuration="7.762990785s" podCreationTimestamp="2025-09-30 15:43:52 +0000 UTC" firstStartedPulling="2025-09-30 15:43:54.68752582 +0000 UTC m=+5056.771126247" lastFinishedPulling="2025-09-30 15:43:58.284108421 +0000 UTC m=+5060.367708848" observedRunningTime="2025-09-30 15:43:59.762790459 +0000 UTC m=+5061.846390906" watchObservedRunningTime="2025-09-30 15:43:59.762990785 +0000 UTC m=+5061.846591212" Sep 30 15:44:02 crc kubenswrapper[4799]: I0930 15:44:02.733554 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:44:02 crc kubenswrapper[4799]: I0930 15:44:02.735312 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:44:02 crc kubenswrapper[4799]: I0930 15:44:02.807819 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:44:12 crc kubenswrapper[4799]: I0930 15:44:12.791455 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:44:12 crc kubenswrapper[4799]: I0930 15:44:12.860604 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wf5h6"] Sep 30 15:44:12 crc kubenswrapper[4799]: I0930 15:44:12.866539 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wf5h6" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerName="registry-server" containerID="cri-o://751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866" gracePeriod=2 Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.437128 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.542584 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-utilities\") pod \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.542861 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvffc\" (UniqueName: \"kubernetes.io/projected/57a77b0b-5d24-4e9d-9a34-aff8d8757873-kube-api-access-zvffc\") pod \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.542934 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-catalog-content\") pod \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\" (UID: \"57a77b0b-5d24-4e9d-9a34-aff8d8757873\") " Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.544486 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-utilities" (OuterVolumeSpecName: "utilities") pod "57a77b0b-5d24-4e9d-9a34-aff8d8757873" (UID: "57a77b0b-5d24-4e9d-9a34-aff8d8757873"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.573393 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a77b0b-5d24-4e9d-9a34-aff8d8757873-kube-api-access-zvffc" (OuterVolumeSpecName: "kube-api-access-zvffc") pod "57a77b0b-5d24-4e9d-9a34-aff8d8757873" (UID: "57a77b0b-5d24-4e9d-9a34-aff8d8757873"). InnerVolumeSpecName "kube-api-access-zvffc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.613585 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a77b0b-5d24-4e9d-9a34-aff8d8757873" (UID: "57a77b0b-5d24-4e9d-9a34-aff8d8757873"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.646167 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvffc\" (UniqueName: \"kubernetes.io/projected/57a77b0b-5d24-4e9d-9a34-aff8d8757873-kube-api-access-zvffc\") on node \"crc\" DevicePath \"\"" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.646232 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.646242 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a77b0b-5d24-4e9d-9a34-aff8d8757873-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.878211 4799 generic.go:334] "Generic (PLEG): container finished" podID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerID="751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866" exitCode=0 Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.878268 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wf5h6" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.878315 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wf5h6" event={"ID":"57a77b0b-5d24-4e9d-9a34-aff8d8757873","Type":"ContainerDied","Data":"751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866"} Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.878581 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wf5h6" event={"ID":"57a77b0b-5d24-4e9d-9a34-aff8d8757873","Type":"ContainerDied","Data":"00cc15b5c7d726db89c4885816ab04afd5447f9a2be37a3669a7df39c3a3d8d7"} Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.878612 4799 scope.go:117] "RemoveContainer" containerID="751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.899486 4799 scope.go:117] "RemoveContainer" containerID="345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.929430 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wf5h6"] Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.936984 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wf5h6"] Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.940361 4799 scope.go:117] "RemoveContainer" containerID="c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.982177 4799 scope.go:117] "RemoveContainer" containerID="751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866" Sep 30 15:44:13 crc kubenswrapper[4799]: E0930 15:44:13.982704 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866\": container with ID starting with 751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866 not found: ID does not exist" containerID="751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.982733 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866"} err="failed to get container status \"751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866\": rpc error: code = NotFound desc = could not find container \"751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866\": container with ID starting with 751334613b3a8e2a17a8daf9a94d99ebf44f4f0a4bf0385dc3e6dae43deaa866 not found: ID does not exist" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.982755 4799 scope.go:117] "RemoveContainer" containerID="345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e" Sep 30 15:44:13 crc kubenswrapper[4799]: E0930 15:44:13.982943 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e\": container with ID starting with 345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e not found: ID does not exist" containerID="345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.982964 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e"} err="failed to get container status \"345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e\": rpc error: code = NotFound desc = could not find container \"345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e\": container with ID starting with 345e8ba2c371446abae3542fe4f7ef4f7b08ca56e1ead2cdcedf42602962d38e not found: ID does not exist" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.982977 4799 scope.go:117] "RemoveContainer" containerID="c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c" Sep 30 15:44:13 crc kubenswrapper[4799]: E0930 15:44:13.983173 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c\": container with ID starting with c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c not found: ID does not exist" containerID="c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c" Sep 30 15:44:13 crc kubenswrapper[4799]: I0930 15:44:13.983193 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c"} err="failed to get container status \"c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c\": rpc error: code = NotFound desc = could not find container \"c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c\": container with ID starting with c6f5f77dfce1de11b96a8ff23938973e8df3fd3931d6bf235b16fe1f45ea630c not found: ID does not exist" Sep 30 15:44:14 crc kubenswrapper[4799]: I0930 15:44:14.531085 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" path="/var/lib/kubelet/pods/57a77b0b-5d24-4e9d-9a34-aff8d8757873/volumes" Sep 30 15:44:16 crc kubenswrapper[4799]: E0930 15:44:16.210205 4799 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.245:47982->38.102.83.245:42733: write tcp 38.102.83.245:47982->38.102.83.245:42733: write: broken pipe Sep 30 15:44:29 crc kubenswrapper[4799]: I0930 15:44:29.648329 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:44:29 crc kubenswrapper[4799]: I0930 15:44:29.649418 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:44:29 crc kubenswrapper[4799]: I0930 15:44:29.649468 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:44:29 crc kubenswrapper[4799]: I0930 15:44:29.650846 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:44:29 crc kubenswrapper[4799]: I0930 15:44:29.650910 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" gracePeriod=600 Sep 30 15:44:30 crc kubenswrapper[4799]: E0930 15:44:30.028753 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:44:30 crc kubenswrapper[4799]: I0930 15:44:30.057702 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" exitCode=0 Sep 30 15:44:30 crc kubenswrapper[4799]: I0930 15:44:30.057749 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109"} Sep 30 15:44:30 crc kubenswrapper[4799]: I0930 15:44:30.057786 4799 scope.go:117] "RemoveContainer" containerID="e4b97db586cc9e0bbe8e063b258329d9e848a6cab8be98493164f05db56b329d" Sep 30 15:44:30 crc kubenswrapper[4799]: I0930 15:44:30.058467 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:44:30 crc kubenswrapper[4799]: E0930 15:44:30.058766 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:44:39 crc kubenswrapper[4799]: E0930 15:44:39.646603 4799 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Sep 30 15:44:41 crc kubenswrapper[4799]: I0930 15:44:41.503669 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:44:41 crc kubenswrapper[4799]: E0930 15:44:41.504344 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:44:56 crc kubenswrapper[4799]: I0930 15:44:56.510216 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:44:56 crc kubenswrapper[4799]: E0930 15:44:56.511491 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.163631 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs"] Sep 30 15:45:00 crc kubenswrapper[4799]: E0930 15:45:00.166164 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerName="extract-utilities" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.166184 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerName="extract-utilities" Sep 30 15:45:00 crc kubenswrapper[4799]: E0930 15:45:00.166208 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerName="registry-server" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.166214 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerName="registry-server" Sep 30 15:45:00 crc kubenswrapper[4799]: E0930 15:45:00.166242 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerName="extract-content" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.166248 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerName="extract-content" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.166796 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a77b0b-5d24-4e9d-9a34-aff8d8757873" containerName="registry-server" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.167811 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.173291 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.175184 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.177758 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs"] Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.293694 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74c49389-2fca-4995-88dc-4bd5aa3ed89e-secret-volume\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.293778 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljbf8\" (UniqueName: \"kubernetes.io/projected/74c49389-2fca-4995-88dc-4bd5aa3ed89e-kube-api-access-ljbf8\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.293970 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74c49389-2fca-4995-88dc-4bd5aa3ed89e-config-volume\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.395482 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74c49389-2fca-4995-88dc-4bd5aa3ed89e-secret-volume\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.395532 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljbf8\" (UniqueName: \"kubernetes.io/projected/74c49389-2fca-4995-88dc-4bd5aa3ed89e-kube-api-access-ljbf8\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.395627 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74c49389-2fca-4995-88dc-4bd5aa3ed89e-config-volume\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.396799 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74c49389-2fca-4995-88dc-4bd5aa3ed89e-config-volume\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.403465 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74c49389-2fca-4995-88dc-4bd5aa3ed89e-secret-volume\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.417722 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljbf8\" (UniqueName: \"kubernetes.io/projected/74c49389-2fca-4995-88dc-4bd5aa3ed89e-kube-api-access-ljbf8\") pod \"collect-profiles-29320785-tvdbs\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.499707 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:00 crc kubenswrapper[4799]: I0930 15:45:00.982749 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs"] Sep 30 15:45:01 crc kubenswrapper[4799]: I0930 15:45:01.418677 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" event={"ID":"74c49389-2fca-4995-88dc-4bd5aa3ed89e","Type":"ContainerStarted","Data":"811eec1604261b0db1407a510dcbb7e7bb54f16d99053fb21ac897486fc767fa"} Sep 30 15:45:01 crc kubenswrapper[4799]: I0930 15:45:01.418741 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" event={"ID":"74c49389-2fca-4995-88dc-4bd5aa3ed89e","Type":"ContainerStarted","Data":"b8d684d29d4de2b1115ed32113934d2e989995fc3ac03030deace939e44dd34f"} Sep 30 15:45:01 crc kubenswrapper[4799]: I0930 15:45:01.441241 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" podStartSLOduration=1.441209052 podStartE2EDuration="1.441209052s" podCreationTimestamp="2025-09-30 15:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:45:01.432816867 +0000 UTC m=+5123.516417304" watchObservedRunningTime="2025-09-30 15:45:01.441209052 +0000 UTC m=+5123.524809479" Sep 30 15:45:02 crc kubenswrapper[4799]: I0930 15:45:02.427923 4799 generic.go:334] "Generic (PLEG): container finished" podID="74c49389-2fca-4995-88dc-4bd5aa3ed89e" containerID="811eec1604261b0db1407a510dcbb7e7bb54f16d99053fb21ac897486fc767fa" exitCode=0 Sep 30 15:45:02 crc kubenswrapper[4799]: I0930 15:45:02.428248 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" event={"ID":"74c49389-2fca-4995-88dc-4bd5aa3ed89e","Type":"ContainerDied","Data":"811eec1604261b0db1407a510dcbb7e7bb54f16d99053fb21ac897486fc767fa"} Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.467846 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" event={"ID":"74c49389-2fca-4995-88dc-4bd5aa3ed89e","Type":"ContainerDied","Data":"b8d684d29d4de2b1115ed32113934d2e989995fc3ac03030deace939e44dd34f"} Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.468383 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8d684d29d4de2b1115ed32113934d2e989995fc3ac03030deace939e44dd34f" Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.474962 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.602391 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74c49389-2fca-4995-88dc-4bd5aa3ed89e-secret-volume\") pod \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.602550 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljbf8\" (UniqueName: \"kubernetes.io/projected/74c49389-2fca-4995-88dc-4bd5aa3ed89e-kube-api-access-ljbf8\") pod \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.602735 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74c49389-2fca-4995-88dc-4bd5aa3ed89e-config-volume\") pod \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\" (UID: \"74c49389-2fca-4995-88dc-4bd5aa3ed89e\") " Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.604898 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74c49389-2fca-4995-88dc-4bd5aa3ed89e-config-volume" (OuterVolumeSpecName: "config-volume") pod "74c49389-2fca-4995-88dc-4bd5aa3ed89e" (UID: "74c49389-2fca-4995-88dc-4bd5aa3ed89e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.608058 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c49389-2fca-4995-88dc-4bd5aa3ed89e-kube-api-access-ljbf8" (OuterVolumeSpecName: "kube-api-access-ljbf8") pod "74c49389-2fca-4995-88dc-4bd5aa3ed89e" (UID: "74c49389-2fca-4995-88dc-4bd5aa3ed89e"). InnerVolumeSpecName "kube-api-access-ljbf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.623081 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c49389-2fca-4995-88dc-4bd5aa3ed89e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "74c49389-2fca-4995-88dc-4bd5aa3ed89e" (UID: "74c49389-2fca-4995-88dc-4bd5aa3ed89e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.705791 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljbf8\" (UniqueName: \"kubernetes.io/projected/74c49389-2fca-4995-88dc-4bd5aa3ed89e-kube-api-access-ljbf8\") on node \"crc\" DevicePath \"\"" Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.705832 4799 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/74c49389-2fca-4995-88dc-4bd5aa3ed89e-config-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:45:04 crc kubenswrapper[4799]: I0930 15:45:04.705843 4799 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/74c49389-2fca-4995-88dc-4bd5aa3ed89e-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 30 15:45:05 crc kubenswrapper[4799]: I0930 15:45:05.476172 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29320785-tvdbs" Sep 30 15:45:05 crc kubenswrapper[4799]: I0930 15:45:05.586089 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd"] Sep 30 15:45:05 crc kubenswrapper[4799]: I0930 15:45:05.595023 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29320740-fksrd"] Sep 30 15:45:06 crc kubenswrapper[4799]: I0930 15:45:06.515593 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06752d82-47d3-4e68-9db7-b20b7d3bc86d" path="/var/lib/kubelet/pods/06752d82-47d3-4e68-9db7-b20b7d3bc86d/volumes" Sep 30 15:45:10 crc kubenswrapper[4799]: I0930 15:45:10.504740 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:45:10 crc kubenswrapper[4799]: E0930 15:45:10.505946 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:45:13 crc kubenswrapper[4799]: I0930 15:45:13.630236 4799 scope.go:117] "RemoveContainer" containerID="6dc59e47406c9988c74d995f8d85982e098d36946dac587fdec43104e35ed64f" Sep 30 15:45:13 crc kubenswrapper[4799]: I0930 15:45:13.690821 4799 scope.go:117] "RemoveContainer" containerID="997a7fd3789451e743f96887058dd1ab8ea90dcb13c08271a8f88c8f0ef9e903" Sep 30 15:45:24 crc kubenswrapper[4799]: I0930 15:45:24.503295 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:45:24 crc kubenswrapper[4799]: E0930 15:45:24.505784 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:45:39 crc kubenswrapper[4799]: I0930 15:45:39.503987 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:45:39 crc kubenswrapper[4799]: E0930 15:45:39.506252 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:45:54 crc kubenswrapper[4799]: I0930 15:45:54.504055 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:45:54 crc kubenswrapper[4799]: E0930 15:45:54.506487 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:46:06 crc kubenswrapper[4799]: I0930 15:46:06.502916 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:46:06 crc kubenswrapper[4799]: E0930 15:46:06.503940 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:46:19 crc kubenswrapper[4799]: I0930 15:46:19.504287 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:46:19 crc kubenswrapper[4799]: E0930 15:46:19.505378 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:46:34 crc kubenswrapper[4799]: I0930 15:46:34.504708 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:46:34 crc kubenswrapper[4799]: E0930 15:46:34.505590 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:46:46 crc kubenswrapper[4799]: I0930 15:46:46.658731 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad0aa86-fab9-465c-9062-00d77763387b" containerID="8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80" exitCode=0 Sep 30 15:46:46 crc kubenswrapper[4799]: I0930 15:46:46.658798 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d2c2l/must-gather-njlvq" event={"ID":"cad0aa86-fab9-465c-9062-00d77763387b","Type":"ContainerDied","Data":"8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80"} Sep 30 15:46:46 crc kubenswrapper[4799]: I0930 15:46:46.660234 4799 scope.go:117] "RemoveContainer" containerID="8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80" Sep 30 15:46:47 crc kubenswrapper[4799]: I0930 15:46:47.126438 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d2c2l_must-gather-njlvq_cad0aa86-fab9-465c-9062-00d77763387b/gather/0.log" Sep 30 15:46:49 crc kubenswrapper[4799]: I0930 15:46:49.507140 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:46:49 crc kubenswrapper[4799]: E0930 15:46:49.508568 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:46:55 crc kubenswrapper[4799]: I0930 15:46:55.652251 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d2c2l/must-gather-njlvq"] Sep 30 15:46:55 crc kubenswrapper[4799]: I0930 15:46:55.653361 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-d2c2l/must-gather-njlvq" podUID="cad0aa86-fab9-465c-9062-00d77763387b" containerName="copy" containerID="cri-o://ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a" gracePeriod=2 Sep 30 15:46:55 crc kubenswrapper[4799]: I0930 15:46:55.668615 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d2c2l/must-gather-njlvq"] Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.238809 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d2c2l_must-gather-njlvq_cad0aa86-fab9-465c-9062-00d77763387b/copy/0.log" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.240100 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.405063 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm5kq\" (UniqueName: \"kubernetes.io/projected/cad0aa86-fab9-465c-9062-00d77763387b-kube-api-access-wm5kq\") pod \"cad0aa86-fab9-465c-9062-00d77763387b\" (UID: \"cad0aa86-fab9-465c-9062-00d77763387b\") " Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.405117 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cad0aa86-fab9-465c-9062-00d77763387b-must-gather-output\") pod \"cad0aa86-fab9-465c-9062-00d77763387b\" (UID: \"cad0aa86-fab9-465c-9062-00d77763387b\") " Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.412640 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cad0aa86-fab9-465c-9062-00d77763387b-kube-api-access-wm5kq" (OuterVolumeSpecName: "kube-api-access-wm5kq") pod "cad0aa86-fab9-465c-9062-00d77763387b" (UID: "cad0aa86-fab9-465c-9062-00d77763387b"). InnerVolumeSpecName "kube-api-access-wm5kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.506803 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm5kq\" (UniqueName: \"kubernetes.io/projected/cad0aa86-fab9-465c-9062-00d77763387b-kube-api-access-wm5kq\") on node \"crc\" DevicePath \"\"" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.607476 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cad0aa86-fab9-465c-9062-00d77763387b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cad0aa86-fab9-465c-9062-00d77763387b" (UID: "cad0aa86-fab9-465c-9062-00d77763387b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.609241 4799 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cad0aa86-fab9-465c-9062-00d77763387b-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.801814 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d2c2l_must-gather-njlvq_cad0aa86-fab9-465c-9062-00d77763387b/copy/0.log" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.803812 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad0aa86-fab9-465c-9062-00d77763387b" containerID="ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a" exitCode=143 Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.803927 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d2c2l/must-gather-njlvq" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.803991 4799 scope.go:117] "RemoveContainer" containerID="ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.846960 4799 scope.go:117] "RemoveContainer" containerID="8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.942477 4799 scope.go:117] "RemoveContainer" containerID="ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a" Sep 30 15:46:56 crc kubenswrapper[4799]: E0930 15:46:56.943087 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a\": container with ID starting with ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a not found: ID does not exist" containerID="ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.943247 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a"} err="failed to get container status \"ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a\": rpc error: code = NotFound desc = could not find container \"ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a\": container with ID starting with ec3a26b0039b4d33bfc67f224f1b3ca61bf4250a0e24e592dc2a3fc23e6c861a not found: ID does not exist" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.943343 4799 scope.go:117] "RemoveContainer" containerID="8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80" Sep 30 15:46:56 crc kubenswrapper[4799]: E0930 15:46:56.943997 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80\": container with ID starting with 8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80 not found: ID does not exist" containerID="8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80" Sep 30 15:46:56 crc kubenswrapper[4799]: I0930 15:46:56.944032 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80"} err="failed to get container status \"8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80\": rpc error: code = NotFound desc = could not find container \"8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80\": container with ID starting with 8d2a6cd177eb0334e3da0f343af1058525a5d66b72547ff3b069da749b4cfe80 not found: ID does not exist" Sep 30 15:46:58 crc kubenswrapper[4799]: I0930 15:46:58.516130 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cad0aa86-fab9-465c-9062-00d77763387b" path="/var/lib/kubelet/pods/cad0aa86-fab9-465c-9062-00d77763387b/volumes" Sep 30 15:47:04 crc kubenswrapper[4799]: I0930 15:47:04.504203 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:47:04 crc kubenswrapper[4799]: E0930 15:47:04.505462 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.913833 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-62wsn"] Sep 30 15:47:16 crc kubenswrapper[4799]: E0930 15:47:16.915425 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cad0aa86-fab9-465c-9062-00d77763387b" containerName="gather" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.915449 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cad0aa86-fab9-465c-9062-00d77763387b" containerName="gather" Sep 30 15:47:16 crc kubenswrapper[4799]: E0930 15:47:16.915481 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cad0aa86-fab9-465c-9062-00d77763387b" containerName="copy" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.915487 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cad0aa86-fab9-465c-9062-00d77763387b" containerName="copy" Sep 30 15:47:16 crc kubenswrapper[4799]: E0930 15:47:16.915523 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c49389-2fca-4995-88dc-4bd5aa3ed89e" containerName="collect-profiles" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.915531 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c49389-2fca-4995-88dc-4bd5aa3ed89e" containerName="collect-profiles" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.915808 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cad0aa86-fab9-465c-9062-00d77763387b" containerName="gather" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.915853 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cad0aa86-fab9-465c-9062-00d77763387b" containerName="copy" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.915871 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c49389-2fca-4995-88dc-4bd5aa3ed89e" containerName="collect-profiles" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.918866 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:16 crc kubenswrapper[4799]: I0930 15:47:16.928065 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62wsn"] Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.014743 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jmt8\" (UniqueName: \"kubernetes.io/projected/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-kube-api-access-8jmt8\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.014806 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-utilities\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.014873 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-catalog-content\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.116543 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jmt8\" (UniqueName: \"kubernetes.io/projected/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-kube-api-access-8jmt8\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.116602 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-utilities\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.116638 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-catalog-content\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.117399 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-catalog-content\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.117443 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-utilities\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.147716 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jmt8\" (UniqueName: \"kubernetes.io/projected/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-kube-api-access-8jmt8\") pod \"community-operators-62wsn\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:17 crc kubenswrapper[4799]: I0930 15:47:17.248875 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:18 crc kubenswrapper[4799]: I0930 15:47:18.042829 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62wsn"] Sep 30 15:47:18 crc kubenswrapper[4799]: I0930 15:47:18.503923 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:47:18 crc kubenswrapper[4799]: E0930 15:47:18.504703 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:47:19 crc kubenswrapper[4799]: I0930 15:47:19.030733 4799 generic.go:334] "Generic (PLEG): container finished" podID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerID="76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e" exitCode=0 Sep 30 15:47:19 crc kubenswrapper[4799]: I0930 15:47:19.030791 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wsn" event={"ID":"ee1eba8e-c7fd-482a-af16-bf9e03bd265f","Type":"ContainerDied","Data":"76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e"} Sep 30 15:47:19 crc kubenswrapper[4799]: I0930 15:47:19.030824 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wsn" event={"ID":"ee1eba8e-c7fd-482a-af16-bf9e03bd265f","Type":"ContainerStarted","Data":"d0cad8ccf79b005e1ea9b805c7f1b9cf6ed9d75821e8cb0505db1a94b1be2fbf"} Sep 30 15:47:23 crc kubenswrapper[4799]: I0930 15:47:23.096762 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wsn" event={"ID":"ee1eba8e-c7fd-482a-af16-bf9e03bd265f","Type":"ContainerStarted","Data":"63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03"} Sep 30 15:47:26 crc kubenswrapper[4799]: I0930 15:47:26.130824 4799 generic.go:334] "Generic (PLEG): container finished" podID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerID="63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03" exitCode=0 Sep 30 15:47:26 crc kubenswrapper[4799]: I0930 15:47:26.131039 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wsn" event={"ID":"ee1eba8e-c7fd-482a-af16-bf9e03bd265f","Type":"ContainerDied","Data":"63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03"} Sep 30 15:47:27 crc kubenswrapper[4799]: I0930 15:47:27.143829 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wsn" event={"ID":"ee1eba8e-c7fd-482a-af16-bf9e03bd265f","Type":"ContainerStarted","Data":"1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770"} Sep 30 15:47:27 crc kubenswrapper[4799]: I0930 15:47:27.175232 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-62wsn" podStartSLOduration=3.424688583 podStartE2EDuration="11.17520415s" podCreationTimestamp="2025-09-30 15:47:16 +0000 UTC" firstStartedPulling="2025-09-30 15:47:19.032397508 +0000 UTC m=+5261.115997935" lastFinishedPulling="2025-09-30 15:47:26.782913075 +0000 UTC m=+5268.866513502" observedRunningTime="2025-09-30 15:47:27.169435322 +0000 UTC m=+5269.253035769" watchObservedRunningTime="2025-09-30 15:47:27.17520415 +0000 UTC m=+5269.258804577" Sep 30 15:47:27 crc kubenswrapper[4799]: I0930 15:47:27.249733 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:27 crc kubenswrapper[4799]: I0930 15:47:27.249804 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:28 crc kubenswrapper[4799]: I0930 15:47:28.300933 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-62wsn" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="registry-server" probeResult="failure" output=< Sep 30 15:47:28 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:47:28 crc kubenswrapper[4799]: > Sep 30 15:47:33 crc kubenswrapper[4799]: I0930 15:47:33.504949 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:47:33 crc kubenswrapper[4799]: E0930 15:47:33.506209 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:47:37 crc kubenswrapper[4799]: I0930 15:47:37.298201 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:37 crc kubenswrapper[4799]: I0930 15:47:37.351998 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:37 crc kubenswrapper[4799]: I0930 15:47:37.549616 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62wsn"] Sep 30 15:47:39 crc kubenswrapper[4799]: I0930 15:47:39.262648 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-62wsn" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="registry-server" containerID="cri-o://1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770" gracePeriod=2 Sep 30 15:47:39 crc kubenswrapper[4799]: I0930 15:47:39.836453 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:39 crc kubenswrapper[4799]: I0930 15:47:39.909005 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-catalog-content\") pod \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " Sep 30 15:47:39 crc kubenswrapper[4799]: I0930 15:47:39.909304 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-utilities\") pod \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " Sep 30 15:47:39 crc kubenswrapper[4799]: I0930 15:47:39.909334 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jmt8\" (UniqueName: \"kubernetes.io/projected/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-kube-api-access-8jmt8\") pod \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\" (UID: \"ee1eba8e-c7fd-482a-af16-bf9e03bd265f\") " Sep 30 15:47:39 crc kubenswrapper[4799]: I0930 15:47:39.910091 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-utilities" (OuterVolumeSpecName: "utilities") pod "ee1eba8e-c7fd-482a-af16-bf9e03bd265f" (UID: "ee1eba8e-c7fd-482a-af16-bf9e03bd265f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:47:39 crc kubenswrapper[4799]: I0930 15:47:39.925145 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-kube-api-access-8jmt8" (OuterVolumeSpecName: "kube-api-access-8jmt8") pod "ee1eba8e-c7fd-482a-af16-bf9e03bd265f" (UID: "ee1eba8e-c7fd-482a-af16-bf9e03bd265f"). InnerVolumeSpecName "kube-api-access-8jmt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:47:39 crc kubenswrapper[4799]: I0930 15:47:39.961615 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee1eba8e-c7fd-482a-af16-bf9e03bd265f" (UID: "ee1eba8e-c7fd-482a-af16-bf9e03bd265f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.011737 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.011825 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.011860 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jmt8\" (UniqueName: \"kubernetes.io/projected/ee1eba8e-c7fd-482a-af16-bf9e03bd265f-kube-api-access-8jmt8\") on node \"crc\" DevicePath \"\"" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.274527 4799 generic.go:334] "Generic (PLEG): container finished" podID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerID="1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770" exitCode=0 Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.274579 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wsn" event={"ID":"ee1eba8e-c7fd-482a-af16-bf9e03bd265f","Type":"ContainerDied","Data":"1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770"} Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.274617 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wsn" event={"ID":"ee1eba8e-c7fd-482a-af16-bf9e03bd265f","Type":"ContainerDied","Data":"d0cad8ccf79b005e1ea9b805c7f1b9cf6ed9d75821e8cb0505db1a94b1be2fbf"} Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.274639 4799 scope.go:117] "RemoveContainer" containerID="1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.274797 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62wsn" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.296492 4799 scope.go:117] "RemoveContainer" containerID="63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.325546 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62wsn"] Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.331841 4799 scope.go:117] "RemoveContainer" containerID="76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.335911 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-62wsn"] Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.372582 4799 scope.go:117] "RemoveContainer" containerID="1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770" Sep 30 15:47:40 crc kubenswrapper[4799]: E0930 15:47:40.373460 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770\": container with ID starting with 1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770 not found: ID does not exist" containerID="1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.373582 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770"} err="failed to get container status \"1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770\": rpc error: code = NotFound desc = could not find container \"1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770\": container with ID starting with 1b3c7482755614471c804c7538f4e0bfc3f3b847def163e5aeda1bb058577770 not found: ID does not exist" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.373733 4799 scope.go:117] "RemoveContainer" containerID="63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03" Sep 30 15:47:40 crc kubenswrapper[4799]: E0930 15:47:40.374036 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03\": container with ID starting with 63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03 not found: ID does not exist" containerID="63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.374111 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03"} err="failed to get container status \"63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03\": rpc error: code = NotFound desc = could not find container \"63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03\": container with ID starting with 63c3b52ea121a822b85ca2b92a8c1b70970b87aa7f50c69c3ddb8118ea0e5b03 not found: ID does not exist" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.374181 4799 scope.go:117] "RemoveContainer" containerID="76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e" Sep 30 15:47:40 crc kubenswrapper[4799]: E0930 15:47:40.374482 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e\": container with ID starting with 76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e not found: ID does not exist" containerID="76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.374514 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e"} err="failed to get container status \"76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e\": rpc error: code = NotFound desc = could not find container \"76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e\": container with ID starting with 76db04ecdf621cdc46e616a26306ef07b55bea64a402677bc915e6fb4cd26b8e not found: ID does not exist" Sep 30 15:47:40 crc kubenswrapper[4799]: I0930 15:47:40.518485 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" path="/var/lib/kubelet/pods/ee1eba8e-c7fd-482a-af16-bf9e03bd265f/volumes" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.069486 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5x6dz/must-gather-hkk9p"] Sep 30 15:47:44 crc kubenswrapper[4799]: E0930 15:47:44.070762 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="extract-content" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.070778 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="extract-content" Sep 30 15:47:44 crc kubenswrapper[4799]: E0930 15:47:44.070797 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="registry-server" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.070803 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="registry-server" Sep 30 15:47:44 crc kubenswrapper[4799]: E0930 15:47:44.070832 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="extract-utilities" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.070839 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="extract-utilities" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.071034 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee1eba8e-c7fd-482a-af16-bf9e03bd265f" containerName="registry-server" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.072259 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.076057 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5x6dz"/"openshift-service-ca.crt" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.077038 4799 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5x6dz"/"kube-root-ca.crt" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.109033 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5x6dz/must-gather-hkk9p"] Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.123041 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24199593-2cd6-401d-9e48-71833da098a8-must-gather-output\") pod \"must-gather-hkk9p\" (UID: \"24199593-2cd6-401d-9e48-71833da098a8\") " pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.123136 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2skr\" (UniqueName: \"kubernetes.io/projected/24199593-2cd6-401d-9e48-71833da098a8-kube-api-access-w2skr\") pod \"must-gather-hkk9p\" (UID: \"24199593-2cd6-401d-9e48-71833da098a8\") " pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.224945 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2skr\" (UniqueName: \"kubernetes.io/projected/24199593-2cd6-401d-9e48-71833da098a8-kube-api-access-w2skr\") pod \"must-gather-hkk9p\" (UID: \"24199593-2cd6-401d-9e48-71833da098a8\") " pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.225111 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24199593-2cd6-401d-9e48-71833da098a8-must-gather-output\") pod \"must-gather-hkk9p\" (UID: \"24199593-2cd6-401d-9e48-71833da098a8\") " pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.225600 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24199593-2cd6-401d-9e48-71833da098a8-must-gather-output\") pod \"must-gather-hkk9p\" (UID: \"24199593-2cd6-401d-9e48-71833da098a8\") " pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.247269 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2skr\" (UniqueName: \"kubernetes.io/projected/24199593-2cd6-401d-9e48-71833da098a8-kube-api-access-w2skr\") pod \"must-gather-hkk9p\" (UID: \"24199593-2cd6-401d-9e48-71833da098a8\") " pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:47:44 crc kubenswrapper[4799]: I0930 15:47:44.407229 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:47:45 crc kubenswrapper[4799]: I0930 15:47:45.084954 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5x6dz/must-gather-hkk9p"] Sep 30 15:47:45 crc kubenswrapper[4799]: I0930 15:47:45.325309 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" event={"ID":"24199593-2cd6-401d-9e48-71833da098a8","Type":"ContainerStarted","Data":"6c7cc3520d4a7b0610ccb5edded05f4929e61bca26871ef18e1c4a716640b6f5"} Sep 30 15:47:46 crc kubenswrapper[4799]: I0930 15:47:46.336889 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" event={"ID":"24199593-2cd6-401d-9e48-71833da098a8","Type":"ContainerStarted","Data":"dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f"} Sep 30 15:47:47 crc kubenswrapper[4799]: I0930 15:47:47.349387 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" event={"ID":"24199593-2cd6-401d-9e48-71833da098a8","Type":"ContainerStarted","Data":"5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1"} Sep 30 15:47:47 crc kubenswrapper[4799]: I0930 15:47:47.368370 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" podStartSLOduration=3.368339249 podStartE2EDuration="3.368339249s" podCreationTimestamp="2025-09-30 15:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:47:47.363805416 +0000 UTC m=+5289.447405863" watchObservedRunningTime="2025-09-30 15:47:47.368339249 +0000 UTC m=+5289.451939716" Sep 30 15:47:47 crc kubenswrapper[4799]: I0930 15:47:47.504898 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:47:47 crc kubenswrapper[4799]: E0930 15:47:47.505171 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.235302 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-lnsdv"] Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.247082 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.249760 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5x6dz"/"default-dockercfg-f4t5g" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.388235 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3055c926-6b88-4eff-8ab8-1a6b1027bf27-host\") pod \"crc-debug-lnsdv\" (UID: \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\") " pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.388398 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wt48\" (UniqueName: \"kubernetes.io/projected/3055c926-6b88-4eff-8ab8-1a6b1027bf27-kube-api-access-9wt48\") pod \"crc-debug-lnsdv\" (UID: \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\") " pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.490292 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wt48\" (UniqueName: \"kubernetes.io/projected/3055c926-6b88-4eff-8ab8-1a6b1027bf27-kube-api-access-9wt48\") pod \"crc-debug-lnsdv\" (UID: \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\") " pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.490413 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3055c926-6b88-4eff-8ab8-1a6b1027bf27-host\") pod \"crc-debug-lnsdv\" (UID: \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\") " pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.490580 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3055c926-6b88-4eff-8ab8-1a6b1027bf27-host\") pod \"crc-debug-lnsdv\" (UID: \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\") " pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.528456 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wt48\" (UniqueName: \"kubernetes.io/projected/3055c926-6b88-4eff-8ab8-1a6b1027bf27-kube-api-access-9wt48\") pod \"crc-debug-lnsdv\" (UID: \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\") " pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:47:50 crc kubenswrapper[4799]: I0930 15:47:50.579849 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:47:51 crc kubenswrapper[4799]: I0930 15:47:51.396866 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" event={"ID":"3055c926-6b88-4eff-8ab8-1a6b1027bf27","Type":"ContainerStarted","Data":"277c5d508716a7ef6d91e96a38a2db2a9fdaa5c6ea0f31162d1bf6d4d16e63ba"} Sep 30 15:47:51 crc kubenswrapper[4799]: I0930 15:47:51.398532 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" event={"ID":"3055c926-6b88-4eff-8ab8-1a6b1027bf27","Type":"ContainerStarted","Data":"f4f9eb8edc8a08b864e0f4615ea1abee8bf531641edb5a000380ad19c25c472b"} Sep 30 15:47:51 crc kubenswrapper[4799]: I0930 15:47:51.418701 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" podStartSLOduration=1.418677 podStartE2EDuration="1.418677s" podCreationTimestamp="2025-09-30 15:47:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:47:51.412211512 +0000 UTC m=+5293.495811939" watchObservedRunningTime="2025-09-30 15:47:51.418677 +0000 UTC m=+5293.502277427" Sep 30 15:47:59 crc kubenswrapper[4799]: I0930 15:47:59.503465 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:47:59 crc kubenswrapper[4799]: E0930 15:47:59.504178 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:48:11 crc kubenswrapper[4799]: I0930 15:48:11.503923 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:48:11 crc kubenswrapper[4799]: E0930 15:48:11.504932 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:48:13 crc kubenswrapper[4799]: I0930 15:48:13.866870 4799 scope.go:117] "RemoveContainer" containerID="976f783642389069ea60d2c6e2a7ba340ee537cb9916c25d5ae6a3ca3692abc0" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.696400 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jbdcl"] Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.699603 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.724823 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbdcl"] Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.818289 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-utilities\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.818357 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-catalog-content\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.818709 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8mnj\" (UniqueName: \"kubernetes.io/projected/2102fbec-7b2f-4ed5-b273-920646fdd1cc-kube-api-access-t8mnj\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.921488 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-utilities\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.921575 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-catalog-content\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.921640 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8mnj\" (UniqueName: \"kubernetes.io/projected/2102fbec-7b2f-4ed5-b273-920646fdd1cc-kube-api-access-t8mnj\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.922039 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-utilities\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.922291 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-catalog-content\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:14 crc kubenswrapper[4799]: I0930 15:48:14.949833 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8mnj\" (UniqueName: \"kubernetes.io/projected/2102fbec-7b2f-4ed5-b273-920646fdd1cc-kube-api-access-t8mnj\") pod \"redhat-operators-jbdcl\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:15 crc kubenswrapper[4799]: I0930 15:48:15.028155 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:15 crc kubenswrapper[4799]: I0930 15:48:15.643238 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbdcl"] Sep 30 15:48:15 crc kubenswrapper[4799]: I0930 15:48:15.717545 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbdcl" event={"ID":"2102fbec-7b2f-4ed5-b273-920646fdd1cc","Type":"ContainerStarted","Data":"7c5c579c32e615e9d44287eca90069201e0d5b5732004e115cb9b7967c0c76bb"} Sep 30 15:48:16 crc kubenswrapper[4799]: I0930 15:48:16.729573 4799 generic.go:334] "Generic (PLEG): container finished" podID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerID="91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6" exitCode=0 Sep 30 15:48:16 crc kubenswrapper[4799]: I0930 15:48:16.729716 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbdcl" event={"ID":"2102fbec-7b2f-4ed5-b273-920646fdd1cc","Type":"ContainerDied","Data":"91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6"} Sep 30 15:48:19 crc kubenswrapper[4799]: I0930 15:48:19.766278 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbdcl" event={"ID":"2102fbec-7b2f-4ed5-b273-920646fdd1cc","Type":"ContainerStarted","Data":"6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe"} Sep 30 15:48:23 crc kubenswrapper[4799]: I0930 15:48:23.503223 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:48:23 crc kubenswrapper[4799]: E0930 15:48:23.504275 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:48:28 crc kubenswrapper[4799]: I0930 15:48:28.889175 4799 generic.go:334] "Generic (PLEG): container finished" podID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerID="6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe" exitCode=0 Sep 30 15:48:28 crc kubenswrapper[4799]: I0930 15:48:28.889257 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbdcl" event={"ID":"2102fbec-7b2f-4ed5-b273-920646fdd1cc","Type":"ContainerDied","Data":"6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe"} Sep 30 15:48:30 crc kubenswrapper[4799]: I0930 15:48:30.946763 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbdcl" event={"ID":"2102fbec-7b2f-4ed5-b273-920646fdd1cc","Type":"ContainerStarted","Data":"36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5"} Sep 30 15:48:30 crc kubenswrapper[4799]: I0930 15:48:30.982385 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jbdcl" podStartSLOduration=4.002257281 podStartE2EDuration="16.982357275s" podCreationTimestamp="2025-09-30 15:48:14 +0000 UTC" firstStartedPulling="2025-09-30 15:48:16.732261306 +0000 UTC m=+5318.815861743" lastFinishedPulling="2025-09-30 15:48:29.71236131 +0000 UTC m=+5331.795961737" observedRunningTime="2025-09-30 15:48:30.97603126 +0000 UTC m=+5333.059631687" watchObservedRunningTime="2025-09-30 15:48:30.982357275 +0000 UTC m=+5333.065957702" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.228347 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hmv88"] Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.232959 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.265558 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hmv88"] Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.289095 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-utilities\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.289188 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqf97\" (UniqueName: \"kubernetes.io/projected/f3cecf17-7641-4ac7-9aa6-d16398036fbb-kube-api-access-gqf97\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.289302 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-catalog-content\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.393829 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-utilities\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.393897 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqf97\" (UniqueName: \"kubernetes.io/projected/f3cecf17-7641-4ac7-9aa6-d16398036fbb-kube-api-access-gqf97\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.393978 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-catalog-content\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.394508 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-catalog-content\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.394734 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-utilities\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.424279 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqf97\" (UniqueName: \"kubernetes.io/projected/f3cecf17-7641-4ac7-9aa6-d16398036fbb-kube-api-access-gqf97\") pod \"redhat-marketplace-hmv88\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.504445 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:48:34 crc kubenswrapper[4799]: E0930 15:48:34.504672 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:48:34 crc kubenswrapper[4799]: I0930 15:48:34.565752 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:35 crc kubenswrapper[4799]: I0930 15:48:35.031951 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:35 crc kubenswrapper[4799]: I0930 15:48:35.032360 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:48:35 crc kubenswrapper[4799]: I0930 15:48:35.622215 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hmv88"] Sep 30 15:48:36 crc kubenswrapper[4799]: I0930 15:48:36.022072 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hmv88" event={"ID":"f3cecf17-7641-4ac7-9aa6-d16398036fbb","Type":"ContainerStarted","Data":"fbf264377268ddcfe4c3dd79e65d17898d22bc377578b3d8c6023a1c243979a8"} Sep 30 15:48:36 crc kubenswrapper[4799]: I0930 15:48:36.233897 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbdcl" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" probeResult="failure" output=< Sep 30 15:48:36 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:48:36 crc kubenswrapper[4799]: > Sep 30 15:48:37 crc kubenswrapper[4799]: I0930 15:48:37.036359 4799 generic.go:334] "Generic (PLEG): container finished" podID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerID="9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093" exitCode=0 Sep 30 15:48:37 crc kubenswrapper[4799]: I0930 15:48:37.037046 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hmv88" event={"ID":"f3cecf17-7641-4ac7-9aa6-d16398036fbb","Type":"ContainerDied","Data":"9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093"} Sep 30 15:48:39 crc kubenswrapper[4799]: I0930 15:48:39.060478 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hmv88" event={"ID":"f3cecf17-7641-4ac7-9aa6-d16398036fbb","Type":"ContainerStarted","Data":"f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895"} Sep 30 15:48:41 crc kubenswrapper[4799]: I0930 15:48:41.085151 4799 generic.go:334] "Generic (PLEG): container finished" podID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerID="f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895" exitCode=0 Sep 30 15:48:41 crc kubenswrapper[4799]: I0930 15:48:41.085798 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hmv88" event={"ID":"f3cecf17-7641-4ac7-9aa6-d16398036fbb","Type":"ContainerDied","Data":"f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895"} Sep 30 15:48:42 crc kubenswrapper[4799]: I0930 15:48:42.102834 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hmv88" event={"ID":"f3cecf17-7641-4ac7-9aa6-d16398036fbb","Type":"ContainerStarted","Data":"a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3"} Sep 30 15:48:42 crc kubenswrapper[4799]: I0930 15:48:42.129539 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hmv88" podStartSLOduration=3.4726566229999998 podStartE2EDuration="8.129517405s" podCreationTimestamp="2025-09-30 15:48:34 +0000 UTC" firstStartedPulling="2025-09-30 15:48:37.067602756 +0000 UTC m=+5339.151203193" lastFinishedPulling="2025-09-30 15:48:41.724463548 +0000 UTC m=+5343.808063975" observedRunningTime="2025-09-30 15:48:42.124959012 +0000 UTC m=+5344.208559429" watchObservedRunningTime="2025-09-30 15:48:42.129517405 +0000 UTC m=+5344.213117832" Sep 30 15:48:44 crc kubenswrapper[4799]: I0930 15:48:44.566031 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:44 crc kubenswrapper[4799]: I0930 15:48:44.566755 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:44 crc kubenswrapper[4799]: I0930 15:48:44.643771 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:46 crc kubenswrapper[4799]: I0930 15:48:46.090101 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbdcl" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" probeResult="failure" output=< Sep 30 15:48:46 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:48:46 crc kubenswrapper[4799]: > Sep 30 15:48:48 crc kubenswrapper[4799]: I0930 15:48:48.514305 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:48:48 crc kubenswrapper[4799]: E0930 15:48:48.515188 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:48:54 crc kubenswrapper[4799]: I0930 15:48:54.632113 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:54 crc kubenswrapper[4799]: I0930 15:48:54.743988 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hmv88"] Sep 30 15:48:55 crc kubenswrapper[4799]: I0930 15:48:55.274841 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hmv88" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerName="registry-server" containerID="cri-o://a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3" gracePeriod=2 Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.078118 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.088856 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbdcl" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" probeResult="failure" output=< Sep 30 15:48:56 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:48:56 crc kubenswrapper[4799]: > Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.200058 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqf97\" (UniqueName: \"kubernetes.io/projected/f3cecf17-7641-4ac7-9aa6-d16398036fbb-kube-api-access-gqf97\") pod \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.200181 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-catalog-content\") pod \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.200284 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-utilities\") pod \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\" (UID: \"f3cecf17-7641-4ac7-9aa6-d16398036fbb\") " Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.201560 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-utilities" (OuterVolumeSpecName: "utilities") pod "f3cecf17-7641-4ac7-9aa6-d16398036fbb" (UID: "f3cecf17-7641-4ac7-9aa6-d16398036fbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.215063 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3cecf17-7641-4ac7-9aa6-d16398036fbb" (UID: "f3cecf17-7641-4ac7-9aa6-d16398036fbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.224929 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3cecf17-7641-4ac7-9aa6-d16398036fbb-kube-api-access-gqf97" (OuterVolumeSpecName: "kube-api-access-gqf97") pod "f3cecf17-7641-4ac7-9aa6-d16398036fbb" (UID: "f3cecf17-7641-4ac7-9aa6-d16398036fbb"). InnerVolumeSpecName "kube-api-access-gqf97". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.294012 4799 generic.go:334] "Generic (PLEG): container finished" podID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerID="a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3" exitCode=0 Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.294513 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hmv88" event={"ID":"f3cecf17-7641-4ac7-9aa6-d16398036fbb","Type":"ContainerDied","Data":"a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3"} Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.294551 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hmv88" event={"ID":"f3cecf17-7641-4ac7-9aa6-d16398036fbb","Type":"ContainerDied","Data":"fbf264377268ddcfe4c3dd79e65d17898d22bc377578b3d8c6023a1c243979a8"} Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.294573 4799 scope.go:117] "RemoveContainer" containerID="a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.294770 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hmv88" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.302946 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.303001 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqf97\" (UniqueName: \"kubernetes.io/projected/f3cecf17-7641-4ac7-9aa6-d16398036fbb-kube-api-access-gqf97\") on node \"crc\" DevicePath \"\"" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.303014 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3cecf17-7641-4ac7-9aa6-d16398036fbb-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.342763 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hmv88"] Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.347463 4799 scope.go:117] "RemoveContainer" containerID="f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.355165 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hmv88"] Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.373939 4799 scope.go:117] "RemoveContainer" containerID="9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.435265 4799 scope.go:117] "RemoveContainer" containerID="a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3" Sep 30 15:48:56 crc kubenswrapper[4799]: E0930 15:48:56.435903 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3\": container with ID starting with a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3 not found: ID does not exist" containerID="a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.435936 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3"} err="failed to get container status \"a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3\": rpc error: code = NotFound desc = could not find container \"a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3\": container with ID starting with a3fdfaf5ed1febc9890042b51952aa1beea6f2bb6d2b94e214a809450f5506c3 not found: ID does not exist" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.435959 4799 scope.go:117] "RemoveContainer" containerID="f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895" Sep 30 15:48:56 crc kubenswrapper[4799]: E0930 15:48:56.437689 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895\": container with ID starting with f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895 not found: ID does not exist" containerID="f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.437715 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895"} err="failed to get container status \"f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895\": rpc error: code = NotFound desc = could not find container \"f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895\": container with ID starting with f925fd15aa57b298a9dd4fa6263b64bd04b6ee5b1142791e079b05340eb01895 not found: ID does not exist" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.437731 4799 scope.go:117] "RemoveContainer" containerID="9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093" Sep 30 15:48:56 crc kubenswrapper[4799]: E0930 15:48:56.440369 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093\": container with ID starting with 9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093 not found: ID does not exist" containerID="9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.440393 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093"} err="failed to get container status \"9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093\": rpc error: code = NotFound desc = could not find container \"9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093\": container with ID starting with 9bf8866439557b221e6e67220f11acb00bcad43e23609e0f9120106887fbd093 not found: ID does not exist" Sep 30 15:48:56 crc kubenswrapper[4799]: I0930 15:48:56.523086 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" path="/var/lib/kubelet/pods/f3cecf17-7641-4ac7-9aa6-d16398036fbb/volumes" Sep 30 15:49:02 crc kubenswrapper[4799]: I0930 15:49:02.505253 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:49:02 crc kubenswrapper[4799]: E0930 15:49:02.508126 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:49:06 crc kubenswrapper[4799]: I0930 15:49:06.085042 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbdcl" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" probeResult="failure" output=< Sep 30 15:49:06 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:49:06 crc kubenswrapper[4799]: > Sep 30 15:49:15 crc kubenswrapper[4799]: I0930 15:49:15.504456 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:49:15 crc kubenswrapper[4799]: E0930 15:49:15.505816 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:49:16 crc kubenswrapper[4799]: I0930 15:49:16.088731 4799 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbdcl" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" probeResult="failure" output=< Sep 30 15:49:16 crc kubenswrapper[4799]: timeout: failed to connect service ":50051" within 1s Sep 30 15:49:16 crc kubenswrapper[4799]: > Sep 30 15:49:25 crc kubenswrapper[4799]: I0930 15:49:25.094455 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:49:25 crc kubenswrapper[4799]: I0930 15:49:25.164228 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:49:25 crc kubenswrapper[4799]: I0930 15:49:25.348072 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbdcl"] Sep 30 15:49:26 crc kubenswrapper[4799]: I0930 15:49:26.664668 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jbdcl" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" containerID="cri-o://36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5" gracePeriod=2 Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.204830 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.373872 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-utilities\") pod \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.374063 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8mnj\" (UniqueName: \"kubernetes.io/projected/2102fbec-7b2f-4ed5-b273-920646fdd1cc-kube-api-access-t8mnj\") pod \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.374376 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-catalog-content\") pod \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\" (UID: \"2102fbec-7b2f-4ed5-b273-920646fdd1cc\") " Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.375551 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-utilities" (OuterVolumeSpecName: "utilities") pod "2102fbec-7b2f-4ed5-b273-920646fdd1cc" (UID: "2102fbec-7b2f-4ed5-b273-920646fdd1cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.388997 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2102fbec-7b2f-4ed5-b273-920646fdd1cc-kube-api-access-t8mnj" (OuterVolumeSpecName: "kube-api-access-t8mnj") pod "2102fbec-7b2f-4ed5-b273-920646fdd1cc" (UID: "2102fbec-7b2f-4ed5-b273-920646fdd1cc"). InnerVolumeSpecName "kube-api-access-t8mnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.476630 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2102fbec-7b2f-4ed5-b273-920646fdd1cc" (UID: "2102fbec-7b2f-4ed5-b273-920646fdd1cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.477872 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.477924 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8mnj\" (UniqueName: \"kubernetes.io/projected/2102fbec-7b2f-4ed5-b273-920646fdd1cc-kube-api-access-t8mnj\") on node \"crc\" DevicePath \"\"" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.477937 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2102fbec-7b2f-4ed5-b273-920646fdd1cc-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.691403 4799 generic.go:334] "Generic (PLEG): container finished" podID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerID="36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5" exitCode=0 Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.691459 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbdcl" event={"ID":"2102fbec-7b2f-4ed5-b273-920646fdd1cc","Type":"ContainerDied","Data":"36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5"} Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.691486 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbdcl" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.691513 4799 scope.go:117] "RemoveContainer" containerID="36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.691496 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbdcl" event={"ID":"2102fbec-7b2f-4ed5-b273-920646fdd1cc","Type":"ContainerDied","Data":"7c5c579c32e615e9d44287eca90069201e0d5b5732004e115cb9b7967c0c76bb"} Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.745030 4799 scope.go:117] "RemoveContainer" containerID="6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.789799 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbdcl"] Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.797999 4799 scope.go:117] "RemoveContainer" containerID="91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.836252 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jbdcl"] Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.894810 4799 scope.go:117] "RemoveContainer" containerID="36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5" Sep 30 15:49:27 crc kubenswrapper[4799]: E0930 15:49:27.902876 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5\": container with ID starting with 36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5 not found: ID does not exist" containerID="36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.902958 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5"} err="failed to get container status \"36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5\": rpc error: code = NotFound desc = could not find container \"36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5\": container with ID starting with 36ae6c300bfccd8f606c2fcd18975d3863e831e41d803d2e4ad69174324eb2a5 not found: ID does not exist" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.903004 4799 scope.go:117] "RemoveContainer" containerID="6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe" Sep 30 15:49:27 crc kubenswrapper[4799]: E0930 15:49:27.906011 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe\": container with ID starting with 6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe not found: ID does not exist" containerID="6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.906103 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe"} err="failed to get container status \"6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe\": rpc error: code = NotFound desc = could not find container \"6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe\": container with ID starting with 6ef270c04b8307f4d1baa09d7fef26911b56fc80433a714f7245666d37f1dabe not found: ID does not exist" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.906170 4799 scope.go:117] "RemoveContainer" containerID="91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6" Sep 30 15:49:27 crc kubenswrapper[4799]: E0930 15:49:27.906571 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6\": container with ID starting with 91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6 not found: ID does not exist" containerID="91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6" Sep 30 15:49:27 crc kubenswrapper[4799]: I0930 15:49:27.906609 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6"} err="failed to get container status \"91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6\": rpc error: code = NotFound desc = could not find container \"91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6\": container with ID starting with 91cc9767e1f4955316105ec8f15323bd2e42dafc2e282129e4fb5b854b2401e6 not found: ID does not exist" Sep 30 15:49:28 crc kubenswrapper[4799]: I0930 15:49:28.539184 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" path="/var/lib/kubelet/pods/2102fbec-7b2f-4ed5-b273-920646fdd1cc/volumes" Sep 30 15:49:29 crc kubenswrapper[4799]: I0930 15:49:29.503834 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:49:29 crc kubenswrapper[4799]: E0930 15:49:29.504127 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:49:31 crc kubenswrapper[4799]: I0930 15:49:31.559444 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84f5689c94-tlpq6_b992c9d0-6ae3-4032-a4d5-f893c4813931/barbican-api/0.log" Sep 30 15:49:31 crc kubenswrapper[4799]: I0930 15:49:31.587233 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84f5689c94-tlpq6_b992c9d0-6ae3-4032-a4d5-f893c4813931/barbican-api-log/0.log" Sep 30 15:49:31 crc kubenswrapper[4799]: I0930 15:49:31.826156 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5888b5f87b-kmzgb_663d1769-2ce0-436c-986e-b07abcb30dca/barbican-keystone-listener/0.log" Sep 30 15:49:31 crc kubenswrapper[4799]: I0930 15:49:31.909065 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5888b5f87b-kmzgb_663d1769-2ce0-436c-986e-b07abcb30dca/barbican-keystone-listener-log/0.log" Sep 30 15:49:32 crc kubenswrapper[4799]: I0930 15:49:32.131192 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7958b68d9f-qk798_e704a0e9-df6f-46c9-a9fd-8e5547031469/barbican-worker/0.log" Sep 30 15:49:32 crc kubenswrapper[4799]: I0930 15:49:32.197757 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7958b68d9f-qk798_e704a0e9-df6f-46c9-a9fd-8e5547031469/barbican-worker-log/0.log" Sep 30 15:49:32 crc kubenswrapper[4799]: I0930 15:49:32.480968 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-l5wjl_4565a6db-10d1-440b-a7b7-dc338c18f70f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:32 crc kubenswrapper[4799]: I0930 15:49:32.886404 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a/ceilometer-notification-agent/0.log" Sep 30 15:49:32 crc kubenswrapper[4799]: I0930 15:49:32.889482 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a/proxy-httpd/0.log" Sep 30 15:49:32 crc kubenswrapper[4799]: I0930 15:49:32.908892 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a/ceilometer-central-agent/0.log" Sep 30 15:49:33 crc kubenswrapper[4799]: I0930 15:49:33.125048 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6dcb5a2b-4d03-4b9f-a92a-4ee965f1502a/sg-core/0.log" Sep 30 15:49:33 crc kubenswrapper[4799]: I0930 15:49:33.227933 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_da52c15f-1f9f-4485-99bc-1ea6b29c94c8/cinder-api/0.log" Sep 30 15:49:33 crc kubenswrapper[4799]: I0930 15:49:33.474600 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_da52c15f-1f9f-4485-99bc-1ea6b29c94c8/cinder-api-log/0.log" Sep 30 15:49:33 crc kubenswrapper[4799]: I0930 15:49:33.690770 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5c94a3cf-a75d-469a-9c13-ca9dd2ff978b/cinder-scheduler/0.log" Sep 30 15:49:33 crc kubenswrapper[4799]: I0930 15:49:33.799905 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5c94a3cf-a75d-469a-9c13-ca9dd2ff978b/probe/0.log" Sep 30 15:49:34 crc kubenswrapper[4799]: I0930 15:49:34.053543 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-gpdgt_2cb70efe-b0f2-4ae1-aa08-c18e9d386a9c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:34 crc kubenswrapper[4799]: I0930 15:49:34.219490 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-llf6t_e2cb3e85-dce2-4274-9472-3d5b10ba4358/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:34 crc kubenswrapper[4799]: I0930 15:49:34.447179 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-q5dph_200d7385-c387-4fc1-b108-fd1873d6da2d/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:34 crc kubenswrapper[4799]: I0930 15:49:34.660208 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-s2gr4_9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd/init/0.log" Sep 30 15:49:34 crc kubenswrapper[4799]: I0930 15:49:34.998082 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-s2gr4_9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd/init/0.log" Sep 30 15:49:35 crc kubenswrapper[4799]: I0930 15:49:35.130228 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-s2gr4_9dc1d581-7bc9-4fb6-8eaf-4835f4df60cd/dnsmasq-dns/0.log" Sep 30 15:49:35 crc kubenswrapper[4799]: I0930 15:49:35.297829 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-svwrv_8da29a07-04e5-4c03-adf7-f642c86abecd/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:35 crc kubenswrapper[4799]: I0930 15:49:35.442225 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0334f39c-a7d6-42f3-a764-04e8e4857c41/glance-httpd/0.log" Sep 30 15:49:35 crc kubenswrapper[4799]: I0930 15:49:35.584157 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0334f39c-a7d6-42f3-a764-04e8e4857c41/glance-log/0.log" Sep 30 15:49:35 crc kubenswrapper[4799]: I0930 15:49:35.684370 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5bbb01bf-9925-4813-94e7-17882aa4d4c5/glance-log/0.log" Sep 30 15:49:35 crc kubenswrapper[4799]: I0930 15:49:35.693762 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5bbb01bf-9925-4813-94e7-17882aa4d4c5/glance-httpd/0.log" Sep 30 15:49:36 crc kubenswrapper[4799]: I0930 15:49:36.113336 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f7c888686-pwv2q_bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6/horizon/3.log" Sep 30 15:49:36 crc kubenswrapper[4799]: I0930 15:49:36.182812 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f7c888686-pwv2q_bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6/horizon/4.log" Sep 30 15:49:36 crc kubenswrapper[4799]: I0930 15:49:36.588182 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-sjpdb_b410520b-9fd5-436d-9f2d-48a9ae6e517a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:36 crc kubenswrapper[4799]: I0930 15:49:36.874319 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7f7c888686-pwv2q_bb8ddcc1-5dc7-4591-9d60-e2192c2a88a6/horizon-log/0.log" Sep 30 15:49:36 crc kubenswrapper[4799]: I0930 15:49:36.896333 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6f7fz_9a20125e-0fe9-4826-b17d-f7b61c73bac3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:37 crc kubenswrapper[4799]: I0930 15:49:37.246063 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29320741-x298c_6f205373-c895-453e-90bd-9c323f6c90ac/keystone-cron/0.log" Sep 30 15:49:37 crc kubenswrapper[4799]: I0930 15:49:37.531392 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-577d747d48-ldwh6_301834e4-5636-4faa-b4b3-2b98f14791a7/keystone-api/0.log" Sep 30 15:49:37 crc kubenswrapper[4799]: I0930 15:49:37.568964 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b4ecfc5d-9019-459f-83e8-8affcb145e78/kube-state-metrics/0.log" Sep 30 15:49:37 crc kubenswrapper[4799]: I0930 15:49:37.857032 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lk8pw_43ae8f90-c778-47e6-8c69-8cdd2b4fa949/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:38 crc kubenswrapper[4799]: I0930 15:49:38.984451 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-67c5b7d44f-24zhg_c013c008-ebe8-4afe-8a76-b9103ae99fa6/neutron-httpd/0.log" Sep 30 15:49:38 crc kubenswrapper[4799]: I0930 15:49:38.987830 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-bqtx6_4ead1575-9ec6-41d9-8eda-f986d3a28713/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:39 crc kubenswrapper[4799]: I0930 15:49:39.368640 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-67c5b7d44f-24zhg_c013c008-ebe8-4afe-8a76-b9103ae99fa6/neutron-api/0.log" Sep 30 15:49:40 crc kubenswrapper[4799]: I0930 15:49:40.503583 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:49:40 crc kubenswrapper[4799]: I0930 15:49:40.546248 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_4f55c7c0-30d5-4349-956a-c06acef9fba0/nova-cell0-conductor-conductor/0.log" Sep 30 15:49:41 crc kubenswrapper[4799]: I0930 15:49:41.672750 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ca946c79-58c1-4f6b-9f61-e685bc6dcda9/nova-cell1-conductor-conductor/0.log" Sep 30 15:49:42 crc kubenswrapper[4799]: I0930 15:49:42.234591 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_28bc13f2-5895-42cf-8ab5-60251eb4d09f/nova-api-api/0.log" Sep 30 15:49:42 crc kubenswrapper[4799]: I0930 15:49:42.825160 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_28bc13f2-5895-42cf-8ab5-60251eb4d09f/nova-api-log/0.log" Sep 30 15:49:42 crc kubenswrapper[4799]: I0930 15:49:42.930433 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"1b05deb58963ab5c0987978afccbf2e463e9a5eb409a2d753bd2741039e41a33"} Sep 30 15:49:43 crc kubenswrapper[4799]: I0930 15:49:43.010717 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_a5fd7be8-ed88-4d3c-b979-8d944ca7b864/nova-cell1-novncproxy-novncproxy/0.log" Sep 30 15:49:43 crc kubenswrapper[4799]: I0930 15:49:43.100090 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-tlstx_2da7cba0-9bbc-47c7-99b0-f204928886fa/nova-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:43 crc kubenswrapper[4799]: I0930 15:49:43.895515 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_fad4220f-11f3-45e3-b5f5-70d210ad5f3f/nova-metadata-log/0.log" Sep 30 15:49:44 crc kubenswrapper[4799]: I0930 15:49:44.591739 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_33a2e46f-eaef-4d54-8af4-ef388d3425dc/mysql-bootstrap/0.log" Sep 30 15:49:44 crc kubenswrapper[4799]: I0930 15:49:44.804276 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e3bd7a13-1fa3-4901-aeb6-3b9011ae330b/memcached/0.log" Sep 30 15:49:44 crc kubenswrapper[4799]: I0930 15:49:44.947228 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_886bc5b6-844e-4a31-8fe2-d62c56108e3a/nova-scheduler-scheduler/0.log" Sep 30 15:49:44 crc kubenswrapper[4799]: I0930 15:49:44.947993 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_33a2e46f-eaef-4d54-8af4-ef388d3425dc/mysql-bootstrap/0.log" Sep 30 15:49:45 crc kubenswrapper[4799]: I0930 15:49:45.187291 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_33a2e46f-eaef-4d54-8af4-ef388d3425dc/galera/0.log" Sep 30 15:49:45 crc kubenswrapper[4799]: I0930 15:49:45.429177 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b86436d8-96e2-44aa-9b4c-b9292a1ca129/mysql-bootstrap/0.log" Sep 30 15:49:45 crc kubenswrapper[4799]: I0930 15:49:45.809336 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b86436d8-96e2-44aa-9b4c-b9292a1ca129/mysql-bootstrap/0.log" Sep 30 15:49:45 crc kubenswrapper[4799]: I0930 15:49:45.886872 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_fad4220f-11f3-45e3-b5f5-70d210ad5f3f/nova-metadata-metadata/0.log" Sep 30 15:49:45 crc kubenswrapper[4799]: I0930 15:49:45.951586 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b86436d8-96e2-44aa-9b4c-b9292a1ca129/galera/0.log" Sep 30 15:49:46 crc kubenswrapper[4799]: I0930 15:49:46.102384 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_93dd541f-bdad-44fb-9450-1baa7acabb40/openstackclient/0.log" Sep 30 15:49:46 crc kubenswrapper[4799]: I0930 15:49:46.271164 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2kvml_cbefa245-5ad5-4303-b166-59574e8a0c65/ovn-controller/0.log" Sep 30 15:49:46 crc kubenswrapper[4799]: I0930 15:49:46.564101 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-r82xm_9ee89b02-9e5e-4c62-ade3-c4a308feefde/openstack-network-exporter/0.log" Sep 30 15:49:46 crc kubenswrapper[4799]: I0930 15:49:46.677238 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v2rs4_17bda665-adad-40d3-9465-46013020af4e/ovsdb-server-init/0.log" Sep 30 15:49:46 crc kubenswrapper[4799]: I0930 15:49:46.931292 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v2rs4_17bda665-adad-40d3-9465-46013020af4e/ovs-vswitchd/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.015778 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v2rs4_17bda665-adad-40d3-9465-46013020af4e/ovsdb-server-init/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.032915 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v2rs4_17bda665-adad-40d3-9465-46013020af4e/ovsdb-server/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.248345 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-499wk_7110fd32-977a-4666-8dc1-d0d217767662/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.293041 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_583080ea-2fdf-461a-8b59-f04c48054eb9/openstack-network-exporter/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.387144 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_583080ea-2fdf-461a-8b59-f04c48054eb9/ovn-northd/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.651162 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c146ba83-511e-4ff2-9a3e-6c83c4c43e9e/ovsdbserver-nb/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.698785 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c146ba83-511e-4ff2-9a3e-6c83c4c43e9e/openstack-network-exporter/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.924310 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8d8fe29c-5270-400a-9fc6-1024482dd54d/openstack-network-exporter/0.log" Sep 30 15:49:47 crc kubenswrapper[4799]: I0930 15:49:47.976568 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8d8fe29c-5270-400a-9fc6-1024482dd54d/ovsdbserver-sb/0.log" Sep 30 15:49:48 crc kubenswrapper[4799]: I0930 15:49:48.299274 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9fd6b988d-fckfn_23d1089b-5d01-44ec-b8de-617485c30c7a/placement-api/0.log" Sep 30 15:49:48 crc kubenswrapper[4799]: I0930 15:49:48.607547 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f6f049b3-7415-4acc-829a-659677059137/setup-container/0.log" Sep 30 15:49:48 crc kubenswrapper[4799]: I0930 15:49:48.661530 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-9fd6b988d-fckfn_23d1089b-5d01-44ec-b8de-617485c30c7a/placement-log/0.log" Sep 30 15:49:48 crc kubenswrapper[4799]: I0930 15:49:48.877538 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f6f049b3-7415-4acc-829a-659677059137/rabbitmq/0.log" Sep 30 15:49:48 crc kubenswrapper[4799]: I0930 15:49:48.896097 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f6f049b3-7415-4acc-829a-659677059137/setup-container/0.log" Sep 30 15:49:48 crc kubenswrapper[4799]: I0930 15:49:48.930810 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2/setup-container/0.log" Sep 30 15:49:49 crc kubenswrapper[4799]: I0930 15:49:49.225470 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2/rabbitmq/0.log" Sep 30 15:49:49 crc kubenswrapper[4799]: I0930 15:49:49.237468 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1bb7f383-fb91-4c6f-8a1d-b284bdf7e6b2/setup-container/0.log" Sep 30 15:49:49 crc kubenswrapper[4799]: I0930 15:49:49.324604 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-4pw5n_aadd9e81-92aa-46fe-bae7-8a0ccddd873f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:49 crc kubenswrapper[4799]: I0930 15:49:49.545038 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-nwf55_286e79d0-072f-422d-82cf-8932a5e9839f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:49 crc kubenswrapper[4799]: I0930 15:49:49.600224 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-h5tt7_2fdc7b08-b5e8-45cb-97d9-d44658909edf/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:49 crc kubenswrapper[4799]: I0930 15:49:49.901579 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-g782p_34ff6ecd-36de-4642-953a-c5d60e8ab31e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:49 crc kubenswrapper[4799]: I0930 15:49:49.920422 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8swsg_5f994a1a-5c13-4a9e-9d44-ea7a8389e943/ssh-known-hosts-edpm-deployment/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.236502 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5cfc9f87d7-nj7t8_f32c852d-173a-4417-a268-38b7a480937e/proxy-server/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.273490 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5cfc9f87d7-nj7t8_f32c852d-173a-4417-a268-38b7a480937e/proxy-httpd/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.511684 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-mbljj_ad219d29-0576-4b77-b5b9-fc2c980b7093/swift-ring-rebalance/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.542637 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/account-reaper/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.619686 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/account-auditor/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.904944 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/account-server/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.935368 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/container-auditor/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.955614 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/account-replicator/0.log" Sep 30 15:49:50 crc kubenswrapper[4799]: I0930 15:49:50.983761 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/container-replicator/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.180474 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/container-server/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.184536 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/container-updater/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.250336 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-auditor/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.341583 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-expirer/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.612205 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-replicator/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.711811 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-server/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.747045 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/rsync/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.771785 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/object-updater/0.log" Sep 30 15:49:51 crc kubenswrapper[4799]: I0930 15:49:51.936521 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1500572d-64d4-483a-a525-bc0530fe4cbb/swift-recon-cron/0.log" Sep 30 15:49:52 crc kubenswrapper[4799]: I0930 15:49:52.260786 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ab210e62-906f-46ad-b80a-fbf00a875124/tempest-tests-tempest-tests-runner/0.log" Sep 30 15:49:52 crc kubenswrapper[4799]: I0930 15:49:52.320036 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-f72vr_a86a2d67-7c0c-44b5-a435-15fccfd80d5e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:49:52 crc kubenswrapper[4799]: I0930 15:49:52.565453 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b3f82270-84d2-409f-b243-cde51bfb251f/test-operator-logs-container/0.log" Sep 30 15:49:52 crc kubenswrapper[4799]: I0930 15:49:52.694756 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9p67d_06f8a2b1-960d-4d41-a2ec-45d1ab5c7f99/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Sep 30 15:50:50 crc kubenswrapper[4799]: I0930 15:50:50.711161 4799 generic.go:334] "Generic (PLEG): container finished" podID="3055c926-6b88-4eff-8ab8-1a6b1027bf27" containerID="277c5d508716a7ef6d91e96a38a2db2a9fdaa5c6ea0f31162d1bf6d4d16e63ba" exitCode=0 Sep 30 15:50:50 crc kubenswrapper[4799]: I0930 15:50:50.711243 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" event={"ID":"3055c926-6b88-4eff-8ab8-1a6b1027bf27","Type":"ContainerDied","Data":"277c5d508716a7ef6d91e96a38a2db2a9fdaa5c6ea0f31162d1bf6d4d16e63ba"} Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.839026 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.860557 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3055c926-6b88-4eff-8ab8-1a6b1027bf27-host\") pod \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\" (UID: \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\") " Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.860704 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3055c926-6b88-4eff-8ab8-1a6b1027bf27-host" (OuterVolumeSpecName: "host") pod "3055c926-6b88-4eff-8ab8-1a6b1027bf27" (UID: "3055c926-6b88-4eff-8ab8-1a6b1027bf27"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.860896 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wt48\" (UniqueName: \"kubernetes.io/projected/3055c926-6b88-4eff-8ab8-1a6b1027bf27-kube-api-access-9wt48\") pod \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\" (UID: \"3055c926-6b88-4eff-8ab8-1a6b1027bf27\") " Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.861374 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3055c926-6b88-4eff-8ab8-1a6b1027bf27-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.902847 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3055c926-6b88-4eff-8ab8-1a6b1027bf27-kube-api-access-9wt48" (OuterVolumeSpecName: "kube-api-access-9wt48") pod "3055c926-6b88-4eff-8ab8-1a6b1027bf27" (UID: "3055c926-6b88-4eff-8ab8-1a6b1027bf27"). InnerVolumeSpecName "kube-api-access-9wt48". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.921599 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-lnsdv"] Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.949220 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-lnsdv"] Sep 30 15:50:51 crc kubenswrapper[4799]: I0930 15:50:51.967325 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wt48\" (UniqueName: \"kubernetes.io/projected/3055c926-6b88-4eff-8ab8-1a6b1027bf27-kube-api-access-9wt48\") on node \"crc\" DevicePath \"\"" Sep 30 15:50:52 crc kubenswrapper[4799]: I0930 15:50:52.519579 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3055c926-6b88-4eff-8ab8-1a6b1027bf27" path="/var/lib/kubelet/pods/3055c926-6b88-4eff-8ab8-1a6b1027bf27/volumes" Sep 30 15:50:52 crc kubenswrapper[4799]: I0930 15:50:52.733706 4799 scope.go:117] "RemoveContainer" containerID="277c5d508716a7ef6d91e96a38a2db2a9fdaa5c6ea0f31162d1bf6d4d16e63ba" Sep 30 15:50:52 crc kubenswrapper[4799]: I0930 15:50:52.733792 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-lnsdv" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.234187 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-k265b"] Sep 30 15:50:53 crc kubenswrapper[4799]: E0930 15:50:53.236935 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerName="registry-server" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.237094 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerName="registry-server" Sep 30 15:50:53 crc kubenswrapper[4799]: E0930 15:50:53.237202 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerName="extract-content" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.237284 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerName="extract-content" Sep 30 15:50:53 crc kubenswrapper[4799]: E0930 15:50:53.237370 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerName="extract-utilities" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.237436 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerName="extract-utilities" Sep 30 15:50:53 crc kubenswrapper[4799]: E0930 15:50:53.237514 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3055c926-6b88-4eff-8ab8-1a6b1027bf27" containerName="container-00" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.237603 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="3055c926-6b88-4eff-8ab8-1a6b1027bf27" containerName="container-00" Sep 30 15:50:53 crc kubenswrapper[4799]: E0930 15:50:53.237769 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="extract-content" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.237858 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="extract-content" Sep 30 15:50:53 crc kubenswrapper[4799]: E0930 15:50:53.237961 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="extract-utilities" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.238043 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="extract-utilities" Sep 30 15:50:53 crc kubenswrapper[4799]: E0930 15:50:53.238135 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.238205 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.238720 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="3055c926-6b88-4eff-8ab8-1a6b1027bf27" containerName="container-00" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.238831 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3cecf17-7641-4ac7-9aa6-d16398036fbb" containerName="registry-server" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.238924 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="2102fbec-7b2f-4ed5-b273-920646fdd1cc" containerName="registry-server" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.240297 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.243544 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5x6dz"/"default-dockercfg-f4t5g" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.294801 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-host\") pod \"crc-debug-k265b\" (UID: \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\") " pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.294897 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl7wc\" (UniqueName: \"kubernetes.io/projected/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-kube-api-access-gl7wc\") pod \"crc-debug-k265b\" (UID: \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\") " pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.396280 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-host\") pod \"crc-debug-k265b\" (UID: \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\") " pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.396366 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl7wc\" (UniqueName: \"kubernetes.io/projected/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-kube-api-access-gl7wc\") pod \"crc-debug-k265b\" (UID: \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\") " pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.396451 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-host\") pod \"crc-debug-k265b\" (UID: \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\") " pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.416327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl7wc\" (UniqueName: \"kubernetes.io/projected/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-kube-api-access-gl7wc\") pod \"crc-debug-k265b\" (UID: \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\") " pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.569515 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:53 crc kubenswrapper[4799]: I0930 15:50:53.746025 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-k265b" event={"ID":"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b","Type":"ContainerStarted","Data":"b8338859d6e79f74d18a4b08dfa137337791d2928a55807d5aea9341efbabc01"} Sep 30 15:50:54 crc kubenswrapper[4799]: I0930 15:50:54.760397 4799 generic.go:334] "Generic (PLEG): container finished" podID="b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b" containerID="fd395305f9d93249756574da42129c05c6666f8b5051d73915f3aedf0e88c1a9" exitCode=0 Sep 30 15:50:54 crc kubenswrapper[4799]: I0930 15:50:54.760507 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-k265b" event={"ID":"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b","Type":"ContainerDied","Data":"fd395305f9d93249756574da42129c05c6666f8b5051d73915f3aedf0e88c1a9"} Sep 30 15:50:55 crc kubenswrapper[4799]: I0930 15:50:55.920710 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:50:55 crc kubenswrapper[4799]: I0930 15:50:55.956067 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl7wc\" (UniqueName: \"kubernetes.io/projected/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-kube-api-access-gl7wc\") pod \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\" (UID: \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\") " Sep 30 15:50:55 crc kubenswrapper[4799]: I0930 15:50:55.956259 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-host\") pod \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\" (UID: \"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b\") " Sep 30 15:50:55 crc kubenswrapper[4799]: I0930 15:50:55.956430 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-host" (OuterVolumeSpecName: "host") pod "b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b" (UID: "b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:50:55 crc kubenswrapper[4799]: I0930 15:50:55.956974 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:50:55 crc kubenswrapper[4799]: I0930 15:50:55.965048 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-kube-api-access-gl7wc" (OuterVolumeSpecName: "kube-api-access-gl7wc") pod "b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b" (UID: "b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b"). InnerVolumeSpecName "kube-api-access-gl7wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:50:56 crc kubenswrapper[4799]: I0930 15:50:56.058695 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl7wc\" (UniqueName: \"kubernetes.io/projected/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b-kube-api-access-gl7wc\") on node \"crc\" DevicePath \"\"" Sep 30 15:50:56 crc kubenswrapper[4799]: I0930 15:50:56.796102 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-k265b" event={"ID":"b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b","Type":"ContainerDied","Data":"b8338859d6e79f74d18a4b08dfa137337791d2928a55807d5aea9341efbabc01"} Sep 30 15:50:56 crc kubenswrapper[4799]: I0930 15:50:56.796457 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8338859d6e79f74d18a4b08dfa137337791d2928a55807d5aea9341efbabc01" Sep 30 15:50:56 crc kubenswrapper[4799]: I0930 15:50:56.796167 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-k265b" Sep 30 15:51:05 crc kubenswrapper[4799]: I0930 15:51:05.071558 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-k265b"] Sep 30 15:51:05 crc kubenswrapper[4799]: I0930 15:51:05.081001 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-k265b"] Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.279223 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-5mvdx"] Sep 30 15:51:06 crc kubenswrapper[4799]: E0930 15:51:06.280272 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b" containerName="container-00" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.280290 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b" containerName="container-00" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.280525 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b" containerName="container-00" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.281433 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.284759 4799 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5x6dz"/"default-dockercfg-f4t5g" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.319856 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w59lq\" (UniqueName: \"kubernetes.io/projected/cdf51621-4a4e-4419-9bc4-e247e487a645-kube-api-access-w59lq\") pod \"crc-debug-5mvdx\" (UID: \"cdf51621-4a4e-4419-9bc4-e247e487a645\") " pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.319987 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdf51621-4a4e-4419-9bc4-e247e487a645-host\") pod \"crc-debug-5mvdx\" (UID: \"cdf51621-4a4e-4419-9bc4-e247e487a645\") " pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.422090 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w59lq\" (UniqueName: \"kubernetes.io/projected/cdf51621-4a4e-4419-9bc4-e247e487a645-kube-api-access-w59lq\") pod \"crc-debug-5mvdx\" (UID: \"cdf51621-4a4e-4419-9bc4-e247e487a645\") " pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.422182 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdf51621-4a4e-4419-9bc4-e247e487a645-host\") pod \"crc-debug-5mvdx\" (UID: \"cdf51621-4a4e-4419-9bc4-e247e487a645\") " pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.422394 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdf51621-4a4e-4419-9bc4-e247e487a645-host\") pod \"crc-debug-5mvdx\" (UID: \"cdf51621-4a4e-4419-9bc4-e247e487a645\") " pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.443780 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w59lq\" (UniqueName: \"kubernetes.io/projected/cdf51621-4a4e-4419-9bc4-e247e487a645-kube-api-access-w59lq\") pod \"crc-debug-5mvdx\" (UID: \"cdf51621-4a4e-4419-9bc4-e247e487a645\") " pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.514834 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b" path="/var/lib/kubelet/pods/b38545e7-9a09-4fe9-b6aa-d32b3dbe4f1b/volumes" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.612010 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.911149 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" event={"ID":"cdf51621-4a4e-4419-9bc4-e247e487a645","Type":"ContainerStarted","Data":"3943af88c04247a00ec13611edb05f2b462fdc7c144cfcf3645c9f93b3cc92e1"} Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.911505 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" event={"ID":"cdf51621-4a4e-4419-9bc4-e247e487a645","Type":"ContainerStarted","Data":"401b9194e7313189974effc5af2b60dd08394aefd28d9728b77bb65e9bbed753"} Sep 30 15:51:06 crc kubenswrapper[4799]: I0930 15:51:06.937606 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" podStartSLOduration=0.937575626 podStartE2EDuration="937.575626ms" podCreationTimestamp="2025-09-30 15:51:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-30 15:51:06.930409377 +0000 UTC m=+5489.014009824" watchObservedRunningTime="2025-09-30 15:51:06.937575626 +0000 UTC m=+5489.021176053" Sep 30 15:51:07 crc kubenswrapper[4799]: I0930 15:51:07.926241 4799 generic.go:334] "Generic (PLEG): container finished" podID="cdf51621-4a4e-4419-9bc4-e247e487a645" containerID="3943af88c04247a00ec13611edb05f2b462fdc7c144cfcf3645c9f93b3cc92e1" exitCode=0 Sep 30 15:51:07 crc kubenswrapper[4799]: I0930 15:51:07.926405 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" event={"ID":"cdf51621-4a4e-4419-9bc4-e247e487a645","Type":"ContainerDied","Data":"3943af88c04247a00ec13611edb05f2b462fdc7c144cfcf3645c9f93b3cc92e1"} Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.040061 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.072290 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-5mvdx"] Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.080186 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5x6dz/crc-debug-5mvdx"] Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.082663 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w59lq\" (UniqueName: \"kubernetes.io/projected/cdf51621-4a4e-4419-9bc4-e247e487a645-kube-api-access-w59lq\") pod \"cdf51621-4a4e-4419-9bc4-e247e487a645\" (UID: \"cdf51621-4a4e-4419-9bc4-e247e487a645\") " Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.083158 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdf51621-4a4e-4419-9bc4-e247e487a645-host\") pod \"cdf51621-4a4e-4419-9bc4-e247e487a645\" (UID: \"cdf51621-4a4e-4419-9bc4-e247e487a645\") " Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.083251 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cdf51621-4a4e-4419-9bc4-e247e487a645-host" (OuterVolumeSpecName: "host") pod "cdf51621-4a4e-4419-9bc4-e247e487a645" (UID: "cdf51621-4a4e-4419-9bc4-e247e487a645"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.092080 4799 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdf51621-4a4e-4419-9bc4-e247e487a645-host\") on node \"crc\" DevicePath \"\"" Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.092117 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf51621-4a4e-4419-9bc4-e247e487a645-kube-api-access-w59lq" (OuterVolumeSpecName: "kube-api-access-w59lq") pod "cdf51621-4a4e-4419-9bc4-e247e487a645" (UID: "cdf51621-4a4e-4419-9bc4-e247e487a645"). InnerVolumeSpecName "kube-api-access-w59lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.193712 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w59lq\" (UniqueName: \"kubernetes.io/projected/cdf51621-4a4e-4419-9bc4-e247e487a645-kube-api-access-w59lq\") on node \"crc\" DevicePath \"\"" Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.945236 4799 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="401b9194e7313189974effc5af2b60dd08394aefd28d9728b77bb65e9bbed753" Sep 30 15:51:09 crc kubenswrapper[4799]: I0930 15:51:09.945263 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/crc-debug-5mvdx" Sep 30 15:51:10 crc kubenswrapper[4799]: I0930 15:51:10.514260 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf51621-4a4e-4419-9bc4-e247e487a645" path="/var/lib/kubelet/pods/cdf51621-4a4e-4419-9bc4-e247e487a645/volumes" Sep 30 15:51:11 crc kubenswrapper[4799]: I0930 15:51:11.001398 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/util/0.log" Sep 30 15:51:11 crc kubenswrapper[4799]: I0930 15:51:11.438104 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/util/0.log" Sep 30 15:51:11 crc kubenswrapper[4799]: I0930 15:51:11.468568 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/pull/0.log" Sep 30 15:51:11 crc kubenswrapper[4799]: I0930 15:51:11.495492 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/pull/0.log" Sep 30 15:51:11 crc kubenswrapper[4799]: I0930 15:51:11.730292 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/util/0.log" Sep 30 15:51:11 crc kubenswrapper[4799]: I0930 15:51:11.731787 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/pull/0.log" Sep 30 15:51:11 crc kubenswrapper[4799]: I0930 15:51:11.785348 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_897656d7d27bbb7e315998e59801eba522b74ade81c27d2ed1cf93939eqgxrd_49fbf469-452c-400c-ba9d-8970fbe1abce/extract/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.002615 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-j7tgf_b8a734fd-7d92-4b2d-9bb4-91eee56047c5/kube-rbac-proxy/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.099145 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-j7tgf_b8a734fd-7d92-4b2d-9bb4-91eee56047c5/manager/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.103520 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kqfcp_e0002170-fb28-4c87-a970-350f92c891bd/kube-rbac-proxy/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.258095 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-kqfcp_e0002170-fb28-4c87-a970-350f92c891bd/manager/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.335914 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-mplqn_8bc24269-6185-41a5-861b-19c88576c223/manager/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.377508 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-mplqn_8bc24269-6185-41a5-861b-19c88576c223/kube-rbac-proxy/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.567707 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-gmxsp_19539cfa-6f83-4636-8193-d2373038f353/kube-rbac-proxy/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.667914 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-gmxsp_19539cfa-6f83-4636-8193-d2373038f353/manager/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.796167 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-8vxq2_08704fb7-1721-4f10-97a1-f2359d01cdaf/kube-rbac-proxy/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.895430 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-8vxq2_08704fb7-1721-4f10-97a1-f2359d01cdaf/manager/0.log" Sep 30 15:51:12 crc kubenswrapper[4799]: I0930 15:51:12.945973 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-k955x_fdb1977f-1341-46ae-a4db-e78e20b6d5bd/kube-rbac-proxy/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.058863 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-k955x_fdb1977f-1341-46ae-a4db-e78e20b6d5bd/manager/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.178121 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-rzjpm_cdca0d27-1d59-4858-85a4-681c1d6abf90/kube-rbac-proxy/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.430185 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d857cc749-rzjpm_cdca0d27-1d59-4858-85a4-681c1d6abf90/manager/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.457018 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-p75vq_2e07068f-3f60-4b35-9465-1adb7cda6525/kube-rbac-proxy/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.497414 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7975b88857-p75vq_2e07068f-3f60-4b35-9465-1adb7cda6525/manager/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.684954 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-vq46v_943ee0ba-2610-4d88-b211-4ddf83cb5b03/kube-rbac-proxy/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.778061 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-vq46v_943ee0ba-2610-4d88-b211-4ddf83cb5b03/manager/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.835204 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-2jnc8_40600429-6eda-47b1-bf61-0e22189f7549/kube-rbac-proxy/0.log" Sep 30 15:51:13 crc kubenswrapper[4799]: I0930 15:51:13.953541 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-2jnc8_40600429-6eda-47b1-bf61-0e22189f7549/manager/0.log" Sep 30 15:51:14 crc kubenswrapper[4799]: I0930 15:51:14.049769 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-pxd4j_c8d8a2c1-6828-49d3-bd9a-baa3fe58790e/kube-rbac-proxy/0.log" Sep 30 15:51:14 crc kubenswrapper[4799]: I0930 15:51:14.133170 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-pxd4j_c8d8a2c1-6828-49d3-bd9a-baa3fe58790e/manager/0.log" Sep 30 15:51:14 crc kubenswrapper[4799]: I0930 15:51:14.261712 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-9mvns_f3e1a985-b655-4dc8-8e5f-f49a6459f512/kube-rbac-proxy/0.log" Sep 30 15:51:14 crc kubenswrapper[4799]: I0930 15:51:14.444585 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64d7b59854-9mvns_f3e1a985-b655-4dc8-8e5f-f49a6459f512/manager/0.log" Sep 30 15:51:14 crc kubenswrapper[4799]: I0930 15:51:14.484942 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-vwjtl_4ecfaaa8-a63c-4f41-a08b-5d497a71e16d/kube-rbac-proxy/0.log" Sep 30 15:51:14 crc kubenswrapper[4799]: I0930 15:51:14.634910 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-c7c776c96-vwjtl_4ecfaaa8-a63c-4f41-a08b-5d497a71e16d/manager/0.log" Sep 30 15:51:14 crc kubenswrapper[4799]: I0930 15:51:14.768252 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-rpsrk_34ffd2cc-7377-4160-b03d-cae022be0262/kube-rbac-proxy/0.log" Sep 30 15:51:14 crc kubenswrapper[4799]: I0930 15:51:14.816834 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-76fcc6dc7c-rpsrk_34ffd2cc-7377-4160-b03d-cae022be0262/manager/0.log" Sep 30 15:51:15 crc kubenswrapper[4799]: I0930 15:51:15.008022 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-q7tzd_68487038-29af-45ea-8619-acd75f3d9a7d/kube-rbac-proxy/0.log" Sep 30 15:51:15 crc kubenswrapper[4799]: I0930 15:51:15.032615 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6d776955-q7tzd_68487038-29af-45ea-8619-acd75f3d9a7d/manager/0.log" Sep 30 15:51:15 crc kubenswrapper[4799]: I0930 15:51:15.219735 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-698bb85c6f-m82xh_28ece3c4-17d4-4b6f-8156-f90cfb794581/kube-rbac-proxy/0.log" Sep 30 15:51:15 crc kubenswrapper[4799]: I0930 15:51:15.274073 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-655f477c67-z5l5q_2cfc3242-481e-4575-9d3a-bbd740b097f6/kube-rbac-proxy/0.log" Sep 30 15:51:15 crc kubenswrapper[4799]: I0930 15:51:15.708922 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-655f477c67-z5l5q_2cfc3242-481e-4575-9d3a-bbd740b097f6/operator/0.log" Sep 30 15:51:15 crc kubenswrapper[4799]: I0930 15:51:15.768833 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-w6tbf_f2bdc984-2b22-4a12-99b0-ef6235691241/registry-server/0.log" Sep 30 15:51:15 crc kubenswrapper[4799]: I0930 15:51:15.996047 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-qrwt4_ee68b34b-53e7-452e-a113-ed35ba5c0b85/kube-rbac-proxy/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.184883 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-qrwt4_ee68b34b-53e7-452e-a113-ed35ba5c0b85/manager/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.320083 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-sz2rp_b9a9cd33-4400-4568-822f-effb32929bb3/kube-rbac-proxy/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.432928 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-sz2rp_b9a9cd33-4400-4568-822f-effb32929bb3/manager/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.528157 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-698bb85c6f-m82xh_28ece3c4-17d4-4b6f-8156-f90cfb794581/manager/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.609663 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-7tlqx_78d18534-fd0c-454f-80f3-b65ddd5a24c9/operator/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.692129 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-68qzx_208f263c-4523-45e3-bf7e-19c8407bc599/kube-rbac-proxy/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.745923 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bc7dc7bd9-68qzx_208f263c-4523-45e3-bf7e-19c8407bc599/manager/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.852756 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-9mwxx_3a3cf54d-3dce-46a0-873c-412159358f6a/kube-rbac-proxy/0.log" Sep 30 15:51:16 crc kubenswrapper[4799]: I0930 15:51:16.943048 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-9mwxx_3a3cf54d-3dce-46a0-873c-412159358f6a/manager/0.log" Sep 30 15:51:17 crc kubenswrapper[4799]: I0930 15:51:17.047568 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-nxbv9_bb413bd9-a2c6-43c9-95d4-c59bbc08e930/kube-rbac-proxy/0.log" Sep 30 15:51:17 crc kubenswrapper[4799]: I0930 15:51:17.172067 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-f66b554c6-nxbv9_bb413bd9-a2c6-43c9-95d4-c59bbc08e930/manager/0.log" Sep 30 15:51:17 crc kubenswrapper[4799]: I0930 15:51:17.173850 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-c7mtd_26c4c0df-9057-493d-9a60-b39c664991a4/kube-rbac-proxy/0.log" Sep 30 15:51:17 crc kubenswrapper[4799]: I0930 15:51:17.224798 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-76669f99c-c7mtd_26c4c0df-9057-493d-9a60-b39c664991a4/manager/0.log" Sep 30 15:51:34 crc kubenswrapper[4799]: I0930 15:51:34.583857 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j5jqc_7c44311c-64e8-4587-95c6-2bdda1c93ad1/control-plane-machine-set-operator/0.log" Sep 30 15:51:34 crc kubenswrapper[4799]: I0930 15:51:34.790700 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jgd4t_018cb267-3656-4202-ba7e-323811533a19/kube-rbac-proxy/0.log" Sep 30 15:51:34 crc kubenswrapper[4799]: I0930 15:51:34.828190 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jgd4t_018cb267-3656-4202-ba7e-323811533a19/machine-api-operator/0.log" Sep 30 15:51:47 crc kubenswrapper[4799]: I0930 15:51:47.343474 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-mn5dh_e3b68d27-f024-4124-99d3-d16f2171c911/cert-manager-controller/0.log" Sep 30 15:51:47 crc kubenswrapper[4799]: I0930 15:51:47.556318 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-vdjtl_9146b875-e4c0-451d-8991-858332cc39d7/cert-manager-cainjector/0.log" Sep 30 15:51:47 crc kubenswrapper[4799]: I0930 15:51:47.704909 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-94zcs_8bb043b3-1966-4b5e-bc13-476e14eb927e/cert-manager-webhook/0.log" Sep 30 15:51:59 crc kubenswrapper[4799]: I0930 15:51:59.648516 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:51:59 crc kubenswrapper[4799]: I0930 15:51:59.649094 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:52:00 crc kubenswrapper[4799]: I0930 15:52:00.518162 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-pbz2f_3c1af7dd-861c-4d2e-8d35-7da2b3ea4a69/nmstate-console-plugin/0.log" Sep 30 15:52:00 crc kubenswrapper[4799]: I0930 15:52:00.742731 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-672fh_60de70ea-f8ce-4f5a-ab69-45fe7cd77e20/nmstate-handler/0.log" Sep 30 15:52:00 crc kubenswrapper[4799]: I0930 15:52:00.846282 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-bk8lx_dc69b189-0b92-42fb-9584-42fd5aed183f/kube-rbac-proxy/0.log" Sep 30 15:52:00 crc kubenswrapper[4799]: I0930 15:52:00.893200 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-bk8lx_dc69b189-0b92-42fb-9584-42fd5aed183f/nmstate-metrics/0.log" Sep 30 15:52:01 crc kubenswrapper[4799]: I0930 15:52:01.005076 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-hh9gx_311ed4a0-c6f5-459d-9da6-86dbe84dd12f/nmstate-operator/0.log" Sep 30 15:52:01 crc kubenswrapper[4799]: I0930 15:52:01.141398 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-jdngw_23e26176-f572-43f8-baba-8bbe27fa0bca/nmstate-webhook/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.007537 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-dcxwg_96370b88-b581-4342-a570-059733689d3e/kube-rbac-proxy/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.022853 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-dcxwg_96370b88-b581-4342-a570-059733689d3e/controller/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.290043 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-frr-files/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.420204 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-frr-files/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.426158 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-metrics/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.427568 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-reloader/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.492821 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-reloader/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.741864 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-metrics/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.765580 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-frr-files/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.772209 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-metrics/0.log" Sep 30 15:52:19 crc kubenswrapper[4799]: I0930 15:52:19.806494 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-reloader/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.011297 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-reloader/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.016030 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-frr-files/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.054761 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/cp-metrics/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.062315 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/controller/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.197708 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/frr-metrics/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.334277 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/kube-rbac-proxy/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.345538 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/kube-rbac-proxy-frr/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.549843 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/reloader/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.613478 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-z8j6r_17444fcf-15e8-4038-9500-508d084a38cd/frr-k8s-webhook-server/0.log" Sep 30 15:52:20 crc kubenswrapper[4799]: I0930 15:52:20.984280 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-668b984c56-cwmvh_72f552b4-b1cb-4659-8629-d8ff13a85cc5/manager/0.log" Sep 30 15:52:21 crc kubenswrapper[4799]: I0930 15:52:21.381799 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-698f89f9d8-jdx5j_624ab48f-5623-4bf1-af7f-4cad0be3e654/webhook-server/0.log" Sep 30 15:52:21 crc kubenswrapper[4799]: I0930 15:52:21.419690 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mtwhr_056d4eef-c8d4-4f76-9006-cb6454ec953c/kube-rbac-proxy/0.log" Sep 30 15:52:21 crc kubenswrapper[4799]: I0930 15:52:21.939086 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-dhdb5_6d15288f-3467-4f02-815d-ce489f8f5ad3/frr/0.log" Sep 30 15:52:22 crc kubenswrapper[4799]: I0930 15:52:22.177762 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mtwhr_056d4eef-c8d4-4f76-9006-cb6454ec953c/speaker/0.log" Sep 30 15:52:29 crc kubenswrapper[4799]: I0930 15:52:29.648461 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:52:29 crc kubenswrapper[4799]: I0930 15:52:29.648838 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:52:34 crc kubenswrapper[4799]: I0930 15:52:34.862462 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/util/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.074586 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/util/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.116786 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/pull/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.176243 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/pull/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.343480 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/util/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.402783 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/pull/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.414415 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bc9f4qh_8bd6c752-45af-4df7-9c3f-dadeeb43bd0c/extract/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.561105 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-utilities/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.783306 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-content/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.831346 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-content/0.log" Sep 30 15:52:35 crc kubenswrapper[4799]: I0930 15:52:35.863163 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-utilities/0.log" Sep 30 15:52:36 crc kubenswrapper[4799]: I0930 15:52:36.067139 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-utilities/0.log" Sep 30 15:52:36 crc kubenswrapper[4799]: I0930 15:52:36.100706 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/extract-content/0.log" Sep 30 15:52:36 crc kubenswrapper[4799]: I0930 15:52:36.354583 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-utilities/0.log" Sep 30 15:52:36 crc kubenswrapper[4799]: I0930 15:52:36.684973 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-content/0.log" Sep 30 15:52:36 crc kubenswrapper[4799]: I0930 15:52:36.692175 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-utilities/0.log" Sep 30 15:52:36 crc kubenswrapper[4799]: I0930 15:52:36.793423 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-content/0.log" Sep 30 15:52:36 crc kubenswrapper[4799]: I0930 15:52:36.976800 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-content/0.log" Sep 30 15:52:37 crc kubenswrapper[4799]: I0930 15:52:37.023445 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/extract-utilities/0.log" Sep 30 15:52:37 crc kubenswrapper[4799]: I0930 15:52:37.255639 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5d6jh_171994f2-bcb6-4253-a9c1-bfac1e7b1332/registry-server/0.log" Sep 30 15:52:37 crc kubenswrapper[4799]: I0930 15:52:37.359060 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/util/0.log" Sep 30 15:52:37 crc kubenswrapper[4799]: I0930 15:52:37.687541 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rnzqz_95e3f792-4726-4292-9aa3-a94919534022/registry-server/0.log" Sep 30 15:52:37 crc kubenswrapper[4799]: I0930 15:52:37.757573 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/pull/0.log" Sep 30 15:52:37 crc kubenswrapper[4799]: I0930 15:52:37.806104 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/util/0.log" Sep 30 15:52:37 crc kubenswrapper[4799]: I0930 15:52:37.852585 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/pull/0.log" Sep 30 15:52:38 crc kubenswrapper[4799]: I0930 15:52:38.251844 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/pull/0.log" Sep 30 15:52:38 crc kubenswrapper[4799]: I0930 15:52:38.270543 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/util/0.log" Sep 30 15:52:38 crc kubenswrapper[4799]: I0930 15:52:38.325455 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96txr2b_90186c62-dccd-4a14-908c-6b6b5c1750ba/extract/0.log" Sep 30 15:52:38 crc kubenswrapper[4799]: I0930 15:52:38.560149 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z2qlr_7f71a790-5180-4e2c-be45-837154fa1d62/marketplace-operator/0.log" Sep 30 15:52:38 crc kubenswrapper[4799]: I0930 15:52:38.586162 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-utilities/0.log" Sep 30 15:52:38 crc kubenswrapper[4799]: I0930 15:52:38.870048 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-utilities/0.log" Sep 30 15:52:38 crc kubenswrapper[4799]: I0930 15:52:38.913135 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-content/0.log" Sep 30 15:52:38 crc kubenswrapper[4799]: I0930 15:52:38.949708 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-content/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.109268 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-utilities/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.134472 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/extract-content/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.307490 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-utilities/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.388010 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rmb2k_edfa00ec-c706-4129-b0d2-baf76f31b44c/registry-server/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.504527 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-utilities/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.520588 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-content/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.593209 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-content/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.761969 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-utilities/0.log" Sep 30 15:52:39 crc kubenswrapper[4799]: I0930 15:52:39.832576 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/extract-content/0.log" Sep 30 15:52:40 crc kubenswrapper[4799]: I0930 15:52:40.021358 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sfssn_de974362-13ef-46e4-9cc2-bc69f6b2db91/registry-server/0.log" Sep 30 15:52:59 crc kubenswrapper[4799]: I0930 15:52:59.648643 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:52:59 crc kubenswrapper[4799]: I0930 15:52:59.649203 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:52:59 crc kubenswrapper[4799]: I0930 15:52:59.649252 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:52:59 crc kubenswrapper[4799]: I0930 15:52:59.650091 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b05deb58963ab5c0987978afccbf2e463e9a5eb409a2d753bd2741039e41a33"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:52:59 crc kubenswrapper[4799]: I0930 15:52:59.650153 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://1b05deb58963ab5c0987978afccbf2e463e9a5eb409a2d753bd2741039e41a33" gracePeriod=600 Sep 30 15:53:00 crc kubenswrapper[4799]: I0930 15:53:00.071218 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="1b05deb58963ab5c0987978afccbf2e463e9a5eb409a2d753bd2741039e41a33" exitCode=0 Sep 30 15:53:00 crc kubenswrapper[4799]: I0930 15:53:00.071467 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"1b05deb58963ab5c0987978afccbf2e463e9a5eb409a2d753bd2741039e41a33"} Sep 30 15:53:00 crc kubenswrapper[4799]: I0930 15:53:00.071785 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerStarted","Data":"2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45"} Sep 30 15:53:00 crc kubenswrapper[4799]: I0930 15:53:00.071824 4799 scope.go:117] "RemoveContainer" containerID="37a487f58692cdfb1acf5051988140494d88c915c2a7b0e5aa8a04aa3363b109" Sep 30 15:53:14 crc kubenswrapper[4799]: E0930 15:53:14.707031 4799 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.245:46954->38.102.83.245:42733: read tcp 38.102.83.245:46954->38.102.83.245:42733: read: connection reset by peer Sep 30 15:55:02 crc kubenswrapper[4799]: I0930 15:55:02.889015 4799 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hsp8b"] Sep 30 15:55:02 crc kubenswrapper[4799]: E0930 15:55:02.890118 4799 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf51621-4a4e-4419-9bc4-e247e487a645" containerName="container-00" Sep 30 15:55:02 crc kubenswrapper[4799]: I0930 15:55:02.890135 4799 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf51621-4a4e-4419-9bc4-e247e487a645" containerName="container-00" Sep 30 15:55:02 crc kubenswrapper[4799]: I0930 15:55:02.890388 4799 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf51621-4a4e-4419-9bc4-e247e487a645" containerName="container-00" Sep 30 15:55:02 crc kubenswrapper[4799]: I0930 15:55:02.896806 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:02 crc kubenswrapper[4799]: I0930 15:55:02.911107 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsp8b"] Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.005165 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89smc\" (UniqueName: \"kubernetes.io/projected/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-kube-api-access-89smc\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.005318 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-catalog-content\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.005427 4799 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-utilities\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.107445 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-utilities\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.107590 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89smc\" (UniqueName: \"kubernetes.io/projected/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-kube-api-access-89smc\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.107636 4799 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-catalog-content\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.108313 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-catalog-content\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.108554 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-utilities\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.129327 4799 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89smc\" (UniqueName: \"kubernetes.io/projected/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-kube-api-access-89smc\") pod \"certified-operators-hsp8b\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.218580 4799 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:03 crc kubenswrapper[4799]: I0930 15:55:03.986841 4799 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsp8b"] Sep 30 15:55:04 crc kubenswrapper[4799]: I0930 15:55:04.481680 4799 generic.go:334] "Generic (PLEG): container finished" podID="1c146de5-6642-4ee2-a1fa-d4d9806adbf6" containerID="e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48" exitCode=0 Sep 30 15:55:04 crc kubenswrapper[4799]: I0930 15:55:04.481733 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsp8b" event={"ID":"1c146de5-6642-4ee2-a1fa-d4d9806adbf6","Type":"ContainerDied","Data":"e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48"} Sep 30 15:55:04 crc kubenswrapper[4799]: I0930 15:55:04.481765 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsp8b" event={"ID":"1c146de5-6642-4ee2-a1fa-d4d9806adbf6","Type":"ContainerStarted","Data":"b07dd1281278e23be8b350c32e9c696354d2394e66244026d970bdcc45b1de9c"} Sep 30 15:55:04 crc kubenswrapper[4799]: I0930 15:55:04.485226 4799 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 30 15:55:05 crc kubenswrapper[4799]: I0930 15:55:05.493904 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsp8b" event={"ID":"1c146de5-6642-4ee2-a1fa-d4d9806adbf6","Type":"ContainerStarted","Data":"7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da"} Sep 30 15:55:06 crc kubenswrapper[4799]: I0930 15:55:06.505916 4799 generic.go:334] "Generic (PLEG): container finished" podID="1c146de5-6642-4ee2-a1fa-d4d9806adbf6" containerID="7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da" exitCode=0 Sep 30 15:55:06 crc kubenswrapper[4799]: I0930 15:55:06.527805 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsp8b" event={"ID":"1c146de5-6642-4ee2-a1fa-d4d9806adbf6","Type":"ContainerDied","Data":"7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da"} Sep 30 15:55:07 crc kubenswrapper[4799]: I0930 15:55:07.521448 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsp8b" event={"ID":"1c146de5-6642-4ee2-a1fa-d4d9806adbf6","Type":"ContainerStarted","Data":"7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d"} Sep 30 15:55:07 crc kubenswrapper[4799]: I0930 15:55:07.555730 4799 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hsp8b" podStartSLOduration=2.955590355 podStartE2EDuration="5.555701529s" podCreationTimestamp="2025-09-30 15:55:02 +0000 UTC" firstStartedPulling="2025-09-30 15:55:04.4849084 +0000 UTC m=+5726.568508827" lastFinishedPulling="2025-09-30 15:55:07.085019574 +0000 UTC m=+5729.168620001" observedRunningTime="2025-09-30 15:55:07.546869381 +0000 UTC m=+5729.630469808" watchObservedRunningTime="2025-09-30 15:55:07.555701529 +0000 UTC m=+5729.639301966" Sep 30 15:55:13 crc kubenswrapper[4799]: I0930 15:55:13.219152 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:13 crc kubenswrapper[4799]: I0930 15:55:13.219852 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:13 crc kubenswrapper[4799]: I0930 15:55:13.279102 4799 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:13 crc kubenswrapper[4799]: I0930 15:55:13.639953 4799 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:13 crc kubenswrapper[4799]: I0930 15:55:13.697624 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hsp8b"] Sep 30 15:55:15 crc kubenswrapper[4799]: I0930 15:55:15.606724 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hsp8b" podUID="1c146de5-6642-4ee2-a1fa-d4d9806adbf6" containerName="registry-server" containerID="cri-o://7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d" gracePeriod=2 Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.077961 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.234741 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-utilities\") pod \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.234838 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-catalog-content\") pod \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.234958 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89smc\" (UniqueName: \"kubernetes.io/projected/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-kube-api-access-89smc\") pod \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\" (UID: \"1c146de5-6642-4ee2-a1fa-d4d9806adbf6\") " Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.235789 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-utilities" (OuterVolumeSpecName: "utilities") pod "1c146de5-6642-4ee2-a1fa-d4d9806adbf6" (UID: "1c146de5-6642-4ee2-a1fa-d4d9806adbf6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.247463 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-kube-api-access-89smc" (OuterVolumeSpecName: "kube-api-access-89smc") pod "1c146de5-6642-4ee2-a1fa-d4d9806adbf6" (UID: "1c146de5-6642-4ee2-a1fa-d4d9806adbf6"). InnerVolumeSpecName "kube-api-access-89smc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.293056 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c146de5-6642-4ee2-a1fa-d4d9806adbf6" (UID: "1c146de5-6642-4ee2-a1fa-d4d9806adbf6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.337893 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89smc\" (UniqueName: \"kubernetes.io/projected/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-kube-api-access-89smc\") on node \"crc\" DevicePath \"\"" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.337931 4799 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-utilities\") on node \"crc\" DevicePath \"\"" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.337944 4799 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c146de5-6642-4ee2-a1fa-d4d9806adbf6-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.621131 4799 generic.go:334] "Generic (PLEG): container finished" podID="1c146de5-6642-4ee2-a1fa-d4d9806adbf6" containerID="7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d" exitCode=0 Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.621189 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsp8b" event={"ID":"1c146de5-6642-4ee2-a1fa-d4d9806adbf6","Type":"ContainerDied","Data":"7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d"} Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.621210 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsp8b" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.621220 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsp8b" event={"ID":"1c146de5-6642-4ee2-a1fa-d4d9806adbf6","Type":"ContainerDied","Data":"b07dd1281278e23be8b350c32e9c696354d2394e66244026d970bdcc45b1de9c"} Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.621240 4799 scope.go:117] "RemoveContainer" containerID="7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.651966 4799 scope.go:117] "RemoveContainer" containerID="7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.666706 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hsp8b"] Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.678901 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hsp8b"] Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.682086 4799 scope.go:117] "RemoveContainer" containerID="e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.723626 4799 scope.go:117] "RemoveContainer" containerID="7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d" Sep 30 15:55:16 crc kubenswrapper[4799]: E0930 15:55:16.724111 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d\": container with ID starting with 7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d not found: ID does not exist" containerID="7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.724157 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d"} err="failed to get container status \"7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d\": rpc error: code = NotFound desc = could not find container \"7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d\": container with ID starting with 7c5e8c2cfa0eb9140ee77ec4320f0d1833f665041413364393d4cea8a7fe983d not found: ID does not exist" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.724184 4799 scope.go:117] "RemoveContainer" containerID="7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da" Sep 30 15:55:16 crc kubenswrapper[4799]: E0930 15:55:16.724567 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da\": container with ID starting with 7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da not found: ID does not exist" containerID="7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.724591 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da"} err="failed to get container status \"7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da\": rpc error: code = NotFound desc = could not find container \"7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da\": container with ID starting with 7a558ee84beaa49f12c27c5abea5d381ace93d0216c0c958678e5cc74c6aa3da not found: ID does not exist" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.724604 4799 scope.go:117] "RemoveContainer" containerID="e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48" Sep 30 15:55:16 crc kubenswrapper[4799]: E0930 15:55:16.724869 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48\": container with ID starting with e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48 not found: ID does not exist" containerID="e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48" Sep 30 15:55:16 crc kubenswrapper[4799]: I0930 15:55:16.724901 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48"} err="failed to get container status \"e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48\": rpc error: code = NotFound desc = could not find container \"e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48\": container with ID starting with e09440c601c7310ca75f2427216d887e60d5909b69c613fd73b36029f5deef48 not found: ID does not exist" Sep 30 15:55:18 crc kubenswrapper[4799]: I0930 15:55:18.517055 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c146de5-6642-4ee2-a1fa-d4d9806adbf6" path="/var/lib/kubelet/pods/1c146de5-6642-4ee2-a1fa-d4d9806adbf6/volumes" Sep 30 15:55:29 crc kubenswrapper[4799]: I0930 15:55:29.650785 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:55:29 crc kubenswrapper[4799]: I0930 15:55:29.651426 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:55:49 crc kubenswrapper[4799]: I0930 15:55:49.974156 4799 generic.go:334] "Generic (PLEG): container finished" podID="24199593-2cd6-401d-9e48-71833da098a8" containerID="dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f" exitCode=0 Sep 30 15:55:49 crc kubenswrapper[4799]: I0930 15:55:49.974284 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" event={"ID":"24199593-2cd6-401d-9e48-71833da098a8","Type":"ContainerDied","Data":"dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f"} Sep 30 15:55:49 crc kubenswrapper[4799]: I0930 15:55:49.975530 4799 scope.go:117] "RemoveContainer" containerID="dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f" Sep 30 15:55:50 crc kubenswrapper[4799]: I0930 15:55:50.389348 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5x6dz_must-gather-hkk9p_24199593-2cd6-401d-9e48-71833da098a8/gather/0.log" Sep 30 15:55:59 crc kubenswrapper[4799]: I0930 15:55:59.648768 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:55:59 crc kubenswrapper[4799]: I0930 15:55:59.649372 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.235510 4799 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5x6dz/must-gather-hkk9p"] Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.237543 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" podUID="24199593-2cd6-401d-9e48-71833da098a8" containerName="copy" containerID="cri-o://5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1" gracePeriod=2 Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.248974 4799 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5x6dz/must-gather-hkk9p"] Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.809850 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5x6dz_must-gather-hkk9p_24199593-2cd6-401d-9e48-71833da098a8/copy/0.log" Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.810423 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.893094 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2skr\" (UniqueName: \"kubernetes.io/projected/24199593-2cd6-401d-9e48-71833da098a8-kube-api-access-w2skr\") pod \"24199593-2cd6-401d-9e48-71833da098a8\" (UID: \"24199593-2cd6-401d-9e48-71833da098a8\") " Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.893468 4799 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24199593-2cd6-401d-9e48-71833da098a8-must-gather-output\") pod \"24199593-2cd6-401d-9e48-71833da098a8\" (UID: \"24199593-2cd6-401d-9e48-71833da098a8\") " Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.900090 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24199593-2cd6-401d-9e48-71833da098a8-kube-api-access-w2skr" (OuterVolumeSpecName: "kube-api-access-w2skr") pod "24199593-2cd6-401d-9e48-71833da098a8" (UID: "24199593-2cd6-401d-9e48-71833da098a8"). InnerVolumeSpecName "kube-api-access-w2skr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 30 15:56:06 crc kubenswrapper[4799]: I0930 15:56:06.997080 4799 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2skr\" (UniqueName: \"kubernetes.io/projected/24199593-2cd6-401d-9e48-71833da098a8-kube-api-access-w2skr\") on node \"crc\" DevicePath \"\"" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.099386 4799 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24199593-2cd6-401d-9e48-71833da098a8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "24199593-2cd6-401d-9e48-71833da098a8" (UID: "24199593-2cd6-401d-9e48-71833da098a8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.100038 4799 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/24199593-2cd6-401d-9e48-71833da098a8-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.162177 4799 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5x6dz_must-gather-hkk9p_24199593-2cd6-401d-9e48-71833da098a8/copy/0.log" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.162828 4799 generic.go:334] "Generic (PLEG): container finished" podID="24199593-2cd6-401d-9e48-71833da098a8" containerID="5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1" exitCode=143 Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.162891 4799 scope.go:117] "RemoveContainer" containerID="5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.162912 4799 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5x6dz/must-gather-hkk9p" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.193817 4799 scope.go:117] "RemoveContainer" containerID="dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.312050 4799 scope.go:117] "RemoveContainer" containerID="5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1" Sep 30 15:56:07 crc kubenswrapper[4799]: E0930 15:56:07.313095 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1\": container with ID starting with 5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1 not found: ID does not exist" containerID="5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.313127 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1"} err="failed to get container status \"5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1\": rpc error: code = NotFound desc = could not find container \"5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1\": container with ID starting with 5149b17b85d2e1396bb16a230a98c24ea548f698a2c40f2f5dde74fe83941fb1 not found: ID does not exist" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.313176 4799 scope.go:117] "RemoveContainer" containerID="dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f" Sep 30 15:56:07 crc kubenswrapper[4799]: E0930 15:56:07.313383 4799 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f\": container with ID starting with dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f not found: ID does not exist" containerID="dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f" Sep 30 15:56:07 crc kubenswrapper[4799]: I0930 15:56:07.313404 4799 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f"} err="failed to get container status \"dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f\": rpc error: code = NotFound desc = could not find container \"dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f\": container with ID starting with dcc00fc0ccc364d29c4c456c33194b847558c5cd8a8626c864e976706baf096f not found: ID does not exist" Sep 30 15:56:08 crc kubenswrapper[4799]: I0930 15:56:08.514618 4799 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24199593-2cd6-401d-9e48-71833da098a8" path="/var/lib/kubelet/pods/24199593-2cd6-401d-9e48-71833da098a8/volumes" Sep 30 15:56:29 crc kubenswrapper[4799]: I0930 15:56:29.648355 4799 patch_prober.go:28] interesting pod/machine-config-daemon-wr2vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 30 15:56:29 crc kubenswrapper[4799]: I0930 15:56:29.650983 4799 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 30 15:56:29 crc kubenswrapper[4799]: I0930 15:56:29.651236 4799 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" Sep 30 15:56:29 crc kubenswrapper[4799]: I0930 15:56:29.653372 4799 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45"} pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 30 15:56:29 crc kubenswrapper[4799]: I0930 15:56:29.653549 4799 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" containerName="machine-config-daemon" containerID="cri-o://2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45" gracePeriod=600 Sep 30 15:56:29 crc kubenswrapper[4799]: E0930 15:56:29.779044 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:56:30 crc kubenswrapper[4799]: I0930 15:56:30.421064 4799 generic.go:334] "Generic (PLEG): container finished" podID="cad3b181-b2f5-4961-8134-f471bff50c40" containerID="2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45" exitCode=0 Sep 30 15:56:30 crc kubenswrapper[4799]: I0930 15:56:30.421153 4799 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" event={"ID":"cad3b181-b2f5-4961-8134-f471bff50c40","Type":"ContainerDied","Data":"2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45"} Sep 30 15:56:30 crc kubenswrapper[4799]: I0930 15:56:30.421200 4799 scope.go:117] "RemoveContainer" containerID="1b05deb58963ab5c0987978afccbf2e463e9a5eb409a2d753bd2741039e41a33" Sep 30 15:56:30 crc kubenswrapper[4799]: I0930 15:56:30.422110 4799 scope.go:117] "RemoveContainer" containerID="2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45" Sep 30 15:56:30 crc kubenswrapper[4799]: E0930 15:56:30.422450 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:56:43 crc kubenswrapper[4799]: I0930 15:56:43.503873 4799 scope.go:117] "RemoveContainer" containerID="2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45" Sep 30 15:56:43 crc kubenswrapper[4799]: E0930 15:56:43.504703 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:56:56 crc kubenswrapper[4799]: I0930 15:56:56.503926 4799 scope.go:117] "RemoveContainer" containerID="2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45" Sep 30 15:56:56 crc kubenswrapper[4799]: E0930 15:56:56.505896 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:57:08 crc kubenswrapper[4799]: I0930 15:57:08.513281 4799 scope.go:117] "RemoveContainer" containerID="2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45" Sep 30 15:57:08 crc kubenswrapper[4799]: E0930 15:57:08.514329 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" Sep 30 15:57:14 crc kubenswrapper[4799]: I0930 15:57:14.252818 4799 scope.go:117] "RemoveContainer" containerID="fd395305f9d93249756574da42129c05c6666f8b5051d73915f3aedf0e88c1a9" Sep 30 15:57:14 crc kubenswrapper[4799]: I0930 15:57:14.278395 4799 scope.go:117] "RemoveContainer" containerID="3943af88c04247a00ec13611edb05f2b462fdc7c144cfcf3645c9f93b3cc92e1" Sep 30 15:57:21 crc kubenswrapper[4799]: I0930 15:57:21.503926 4799 scope.go:117] "RemoveContainer" containerID="2ccb858025c7a6ddd6ccbedc205583e461f0f6aff2c30d24ccb49126969c9e45" Sep 30 15:57:21 crc kubenswrapper[4799]: E0930 15:57:21.504624 4799 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-wr2vb_openshift-machine-config-operator(cad3b181-b2f5-4961-8134-f471bff50c40)\"" pod="openshift-machine-config-operator/machine-config-daemon-wr2vb" podUID="cad3b181-b2f5-4961-8134-f471bff50c40" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515066776754024474 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015066776755017412 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015066762753016526 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015066762754015477 5ustar corecore